var/home/core/zuul-output/0000755000175000017500000000000015071307225014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071334517015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006511663315071334510017707 0ustar rootrootOct 07 22:08:45 crc systemd[1]: Starting Kubernetes Kubelet... Oct 07 22:08:45 crc restorecon[4723]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:45 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 22:08:46 crc restorecon[4723]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 07 22:08:46 crc kubenswrapper[4871]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 22:08:46 crc kubenswrapper[4871]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 07 22:08:46 crc kubenswrapper[4871]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 22:08:46 crc kubenswrapper[4871]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 22:08:46 crc kubenswrapper[4871]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 07 22:08:46 crc kubenswrapper[4871]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.676710 4871 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685136 4871 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685208 4871 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685217 4871 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685227 4871 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685235 4871 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685245 4871 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685256 4871 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685265 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685275 4871 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685287 4871 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685297 4871 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685307 4871 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685319 4871 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685332 4871 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685343 4871 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685353 4871 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685361 4871 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685368 4871 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685375 4871 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685383 4871 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685392 4871 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685401 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685409 4871 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685417 4871 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685425 4871 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685432 4871 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685440 4871 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685448 4871 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685455 4871 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685463 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685471 4871 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685479 4871 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685486 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685494 4871 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685501 4871 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685509 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685517 4871 feature_gate.go:330] unrecognized feature gate: Example Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685524 4871 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685532 4871 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685541 4871 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685550 4871 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685557 4871 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685564 4871 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685573 4871 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685583 4871 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685591 4871 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685600 4871 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685608 4871 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685616 4871 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685623 4871 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685631 4871 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685638 4871 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685646 4871 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685653 4871 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685661 4871 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685668 4871 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685676 4871 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685684 4871 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685695 4871 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685703 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685710 4871 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685720 4871 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685730 4871 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685738 4871 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685746 4871 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685754 4871 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685762 4871 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685769 4871 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685776 4871 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.685784 4871 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.686012 4871 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687269 4871 flags.go:64] FLAG: --address="0.0.0.0" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687301 4871 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687318 4871 flags.go:64] FLAG: --anonymous-auth="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687332 4871 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687355 4871 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687366 4871 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687386 4871 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687398 4871 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687408 4871 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687419 4871 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687432 4871 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687445 4871 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687456 4871 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687467 4871 flags.go:64] FLAG: --cgroup-root="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687478 4871 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687489 4871 flags.go:64] FLAG: --client-ca-file="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687500 4871 flags.go:64] FLAG: --cloud-config="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687511 4871 flags.go:64] FLAG: --cloud-provider="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687522 4871 flags.go:64] FLAG: --cluster-dns="[]" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687539 4871 flags.go:64] FLAG: --cluster-domain="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687550 4871 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687561 4871 flags.go:64] FLAG: --config-dir="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687572 4871 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687586 4871 flags.go:64] FLAG: --container-log-max-files="5" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687600 4871 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687611 4871 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687622 4871 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687636 4871 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687647 4871 flags.go:64] FLAG: --contention-profiling="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687659 4871 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687669 4871 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687681 4871 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687692 4871 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687707 4871 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687718 4871 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687729 4871 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687739 4871 flags.go:64] FLAG: --enable-load-reader="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687750 4871 flags.go:64] FLAG: --enable-server="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687761 4871 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687776 4871 flags.go:64] FLAG: --event-burst="100" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687788 4871 flags.go:64] FLAG: --event-qps="50" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687834 4871 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687846 4871 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687857 4871 flags.go:64] FLAG: --eviction-hard="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687873 4871 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687884 4871 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687895 4871 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687907 4871 flags.go:64] FLAG: --eviction-soft="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687918 4871 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687929 4871 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687940 4871 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687952 4871 flags.go:64] FLAG: --experimental-mounter-path="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687963 4871 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687974 4871 flags.go:64] FLAG: --fail-swap-on="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687985 4871 flags.go:64] FLAG: --feature-gates="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.687998 4871 flags.go:64] FLAG: --file-check-frequency="20s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688010 4871 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688019 4871 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688029 4871 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688040 4871 flags.go:64] FLAG: --healthz-port="10248" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688050 4871 flags.go:64] FLAG: --help="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688059 4871 flags.go:64] FLAG: --hostname-override="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688067 4871 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688076 4871 flags.go:64] FLAG: --http-check-frequency="20s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688085 4871 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688094 4871 flags.go:64] FLAG: --image-credential-provider-config="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688104 4871 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688113 4871 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688121 4871 flags.go:64] FLAG: --image-service-endpoint="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688130 4871 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688138 4871 flags.go:64] FLAG: --kube-api-burst="100" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688149 4871 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688159 4871 flags.go:64] FLAG: --kube-api-qps="50" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688169 4871 flags.go:64] FLAG: --kube-reserved="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688178 4871 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688187 4871 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688197 4871 flags.go:64] FLAG: --kubelet-cgroups="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688206 4871 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688215 4871 flags.go:64] FLAG: --lock-file="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688224 4871 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688233 4871 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688242 4871 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688255 4871 flags.go:64] FLAG: --log-json-split-stream="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688264 4871 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688272 4871 flags.go:64] FLAG: --log-text-split-stream="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688281 4871 flags.go:64] FLAG: --logging-format="text" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688291 4871 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688303 4871 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688314 4871 flags.go:64] FLAG: --manifest-url="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688324 4871 flags.go:64] FLAG: --manifest-url-header="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688339 4871 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688351 4871 flags.go:64] FLAG: --max-open-files="1000000" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688365 4871 flags.go:64] FLAG: --max-pods="110" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688374 4871 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688383 4871 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688394 4871 flags.go:64] FLAG: --memory-manager-policy="None" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688403 4871 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688414 4871 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688426 4871 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688437 4871 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688464 4871 flags.go:64] FLAG: --node-status-max-images="50" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688475 4871 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688489 4871 flags.go:64] FLAG: --oom-score-adj="-999" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688583 4871 flags.go:64] FLAG: --pod-cidr="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688597 4871 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688616 4871 flags.go:64] FLAG: --pod-manifest-path="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688625 4871 flags.go:64] FLAG: --pod-max-pids="-1" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688638 4871 flags.go:64] FLAG: --pods-per-core="0" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688649 4871 flags.go:64] FLAG: --port="10250" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688660 4871 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688671 4871 flags.go:64] FLAG: --provider-id="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688682 4871 flags.go:64] FLAG: --qos-reserved="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688693 4871 flags.go:64] FLAG: --read-only-port="10255" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688704 4871 flags.go:64] FLAG: --register-node="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688716 4871 flags.go:64] FLAG: --register-schedulable="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688727 4871 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688781 4871 flags.go:64] FLAG: --registry-burst="10" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688827 4871 flags.go:64] FLAG: --registry-qps="5" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688839 4871 flags.go:64] FLAG: --reserved-cpus="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688850 4871 flags.go:64] FLAG: --reserved-memory="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688865 4871 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688877 4871 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688888 4871 flags.go:64] FLAG: --rotate-certificates="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688901 4871 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688913 4871 flags.go:64] FLAG: --runonce="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688924 4871 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688939 4871 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688950 4871 flags.go:64] FLAG: --seccomp-default="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688961 4871 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688971 4871 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688983 4871 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.688997 4871 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689008 4871 flags.go:64] FLAG: --storage-driver-password="root" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689018 4871 flags.go:64] FLAG: --storage-driver-secure="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689029 4871 flags.go:64] FLAG: --storage-driver-table="stats" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689041 4871 flags.go:64] FLAG: --storage-driver-user="root" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689052 4871 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689064 4871 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689076 4871 flags.go:64] FLAG: --system-cgroups="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689087 4871 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689105 4871 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689125 4871 flags.go:64] FLAG: --tls-cert-file="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689136 4871 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689151 4871 flags.go:64] FLAG: --tls-min-version="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689162 4871 flags.go:64] FLAG: --tls-private-key-file="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689173 4871 flags.go:64] FLAG: --topology-manager-policy="none" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689184 4871 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689196 4871 flags.go:64] FLAG: --topology-manager-scope="container" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689208 4871 flags.go:64] FLAG: --v="2" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689222 4871 flags.go:64] FLAG: --version="false" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689237 4871 flags.go:64] FLAG: --vmodule="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689250 4871 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.689263 4871 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689572 4871 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689588 4871 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689597 4871 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689605 4871 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689614 4871 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689622 4871 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689630 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689638 4871 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689645 4871 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689653 4871 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689660 4871 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689668 4871 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689675 4871 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689683 4871 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689692 4871 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689700 4871 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689708 4871 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689716 4871 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689724 4871 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689732 4871 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689746 4871 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689754 4871 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689764 4871 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689775 4871 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689785 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689822 4871 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689831 4871 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689841 4871 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689851 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689861 4871 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689871 4871 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689880 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689890 4871 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689899 4871 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689909 4871 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689919 4871 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689928 4871 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689936 4871 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689944 4871 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689951 4871 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689959 4871 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689969 4871 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689978 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689986 4871 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.689993 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690000 4871 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690016 4871 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690024 4871 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690031 4871 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690039 4871 feature_gate.go:330] unrecognized feature gate: Example Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690049 4871 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690057 4871 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690067 4871 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690075 4871 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690084 4871 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690094 4871 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690104 4871 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690112 4871 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690120 4871 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690129 4871 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690137 4871 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690145 4871 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690154 4871 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690161 4871 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690169 4871 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690176 4871 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690184 4871 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690191 4871 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690199 4871 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690206 4871 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.690214 4871 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.690241 4871 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.705421 4871 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.705471 4871 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705584 4871 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705593 4871 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705597 4871 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705604 4871 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705611 4871 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705615 4871 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705620 4871 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705624 4871 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705630 4871 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705635 4871 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705639 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705643 4871 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705647 4871 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705650 4871 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705655 4871 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705660 4871 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705666 4871 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705671 4871 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705676 4871 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705681 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705686 4871 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705690 4871 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705695 4871 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705700 4871 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705704 4871 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705709 4871 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705713 4871 feature_gate.go:330] unrecognized feature gate: Example Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705719 4871 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705724 4871 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705728 4871 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705734 4871 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705740 4871 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705745 4871 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705750 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705754 4871 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705758 4871 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705763 4871 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705767 4871 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705771 4871 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705776 4871 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705780 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705784 4871 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705806 4871 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705810 4871 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705815 4871 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705818 4871 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705823 4871 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705828 4871 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705832 4871 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705837 4871 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705841 4871 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705846 4871 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705851 4871 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705855 4871 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705859 4871 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705862 4871 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705866 4871 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705869 4871 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705873 4871 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705876 4871 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705880 4871 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705883 4871 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705888 4871 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705891 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705895 4871 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705901 4871 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705904 4871 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705908 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705911 4871 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705916 4871 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.705919 4871 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.705928 4871 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706083 4871 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706094 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706100 4871 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706105 4871 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706109 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706117 4871 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706123 4871 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706130 4871 feature_gate.go:330] unrecognized feature gate: Example Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706136 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706142 4871 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706147 4871 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706152 4871 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706157 4871 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706162 4871 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706166 4871 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706170 4871 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706174 4871 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706178 4871 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706182 4871 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706186 4871 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706189 4871 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706193 4871 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706197 4871 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706201 4871 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706205 4871 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706209 4871 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706212 4871 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706216 4871 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706220 4871 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706224 4871 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706228 4871 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706231 4871 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706235 4871 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706239 4871 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706263 4871 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706268 4871 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706272 4871 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706277 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706281 4871 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706286 4871 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706290 4871 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706294 4871 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706299 4871 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706303 4871 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706307 4871 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706312 4871 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706316 4871 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706319 4871 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706324 4871 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706329 4871 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706334 4871 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706338 4871 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706343 4871 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706371 4871 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706377 4871 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706380 4871 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706384 4871 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706388 4871 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706393 4871 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706397 4871 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706403 4871 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706407 4871 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706411 4871 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706415 4871 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706419 4871 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706423 4871 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706428 4871 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706432 4871 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706436 4871 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706440 4871 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.706445 4871 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.706450 4871 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.706736 4871 server.go:940] "Client rotation is on, will bootstrap in background" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.710905 4871 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.711004 4871 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.712819 4871 server.go:997] "Starting client certificate rotation" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.712842 4871 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.712996 4871 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-19 04:36:26.46817185 +0000 UTC Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.713114 4871 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1014h27m39.755061598s for next certificate rotation Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.738504 4871 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.741537 4871 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.760153 4871 log.go:25] "Validated CRI v1 runtime API" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.803235 4871 log.go:25] "Validated CRI v1 image API" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.806314 4871 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.815825 4871 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-07-22-03-58-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.815895 4871 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.853476 4871 manager.go:217] Machine: {Timestamp:2025-10-07 22:08:46.847353232 +0000 UTC m=+0.650051365 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:32db1e84-0947-42a3-be43-1cbd898711cc BootID:78aa46f6-3267-4ce5-ade1-72b92c4bf3c8 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:93:99:00 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:93:99:00 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:6a:d6:03 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:b8:fc:5f Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:23:0d:0f Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:17:bd:1e Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:47:fb:4d Speed:-1 Mtu:1496} {Name:eth10 MacAddress:32:56:1c:ac:3d:e5 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:4a:1f:2f:e6:ae:76 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.853963 4871 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.854191 4871 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.855147 4871 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.855525 4871 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.855598 4871 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.856036 4871 topology_manager.go:138] "Creating topology manager with none policy" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.856055 4871 container_manager_linux.go:303] "Creating device plugin manager" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.857273 4871 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.857330 4871 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.858309 4871 state_mem.go:36] "Initialized new in-memory state store" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.858466 4871 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.862933 4871 kubelet.go:418] "Attempting to sync node with API server" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.862973 4871 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.863033 4871 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.863059 4871 kubelet.go:324] "Adding apiserver pod source" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.863080 4871 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.869451 4871 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.871127 4871 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.871826 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:46 crc kubenswrapper[4871]: E1007 22:08:46.871959 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.871828 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:46 crc kubenswrapper[4871]: E1007 22:08:46.872055 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.874631 4871 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.876938 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877023 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877046 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877068 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877096 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877115 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877136 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877168 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877194 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877214 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877243 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.877262 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.878313 4871 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.879511 4871 server.go:1280] "Started kubelet" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.880677 4871 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.881840 4871 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.881845 4871 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.882197 4871 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.882246 4871 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.882644 4871 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.882683 4871 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 07:58:28.133829763 +0000 UTC Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.882784 4871 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1281h49m41.251050847s for next certificate rotation Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.882865 4871 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.882879 4871 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 07 22:08:46 crc kubenswrapper[4871]: E1007 22:08:46.882919 4871 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 07 22:08:46 crc systemd[1]: Started Kubernetes Kubelet. Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.883146 4871 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 07 22:08:46 crc kubenswrapper[4871]: E1007 22:08:46.886226 4871 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="200ms" Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.887008 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:46 crc kubenswrapper[4871]: E1007 22:08:46.887160 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.892407 4871 factory.go:55] Registering systemd factory Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.892472 4871 factory.go:221] Registration of the systemd container factory successfully Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.893416 4871 factory.go:153] Registering CRI-O factory Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.893487 4871 factory.go:221] Registration of the crio container factory successfully Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.894330 4871 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.894625 4871 factory.go:103] Registering Raw factory Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.895690 4871 manager.go:1196] Started watching for new ooms in manager Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.905106 4871 manager.go:319] Starting recovery of all containers Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.906105 4871 server.go:460] "Adding debug handlers to kubelet server" Oct 07 22:08:46 crc kubenswrapper[4871]: E1007 22:08:46.905192 4871 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.68:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c54ebf9c3e9d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-07 22:08:46.879435219 +0000 UTC m=+0.682133342,LastTimestamp:2025-10-07 22:08:46.879435219 +0000 UTC m=+0.682133342,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917148 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917234 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917261 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917281 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917305 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917328 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917350 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917373 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917399 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917423 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917446 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917471 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917496 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917522 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917547 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917565 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917586 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917607 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917626 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917649 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917673 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917696 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917715 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917734 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917758 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917776 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917828 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917854 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917873 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917892 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917910 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917929 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917947 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917967 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.917989 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918009 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918027 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918047 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918066 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918086 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918108 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918127 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918146 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918164 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918194 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918215 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918233 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918251 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918273 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.918294 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.922714 4871 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.922854 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.922904 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.922946 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.922977 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923008 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923036 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923060 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923084 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923107 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923131 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923154 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923190 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923215 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923248 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923277 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923302 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923325 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923350 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923375 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923397 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923421 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923445 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923467 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923498 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923520 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923543 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923564 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923587 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923610 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923635 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923657 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923679 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923700 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923722 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923743 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923768 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923818 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923842 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923867 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923894 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923917 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923962 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.923985 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924010 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924032 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924056 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924622 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924679 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924710 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924741 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924778 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924862 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924895 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924926 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.924969 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925007 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925032 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925105 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925447 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925475 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925497 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925525 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925549 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925571 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925591 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925611 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925633 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925651 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925671 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925690 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925725 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925758 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925783 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925831 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925851 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925872 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925898 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925929 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925950 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925976 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.925996 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926016 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926036 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926056 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926078 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926097 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926127 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926155 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926174 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926193 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926213 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926238 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926256 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926276 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926301 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926330 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926363 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926390 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926418 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926448 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926478 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926497 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926516 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926537 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926554 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926579 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926600 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926626 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926647 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926676 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926702 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926719 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926742 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926760 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926780 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926839 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926859 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926888 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926919 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.926991 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927019 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927048 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927076 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927094 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927122 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927141 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927160 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927178 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927205 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927234 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927254 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927280 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927300 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927318 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927336 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927358 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927378 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927397 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927424 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927442 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927462 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927487 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927505 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927523 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927541 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927558 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927577 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927595 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927615 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927634 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927653 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927670 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927690 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927719 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927738 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927765 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927858 4871 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927882 4871 reconstruct.go:97] "Volume reconstruction finished" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.927897 4871 reconciler.go:26] "Reconciler: start to sync state" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.931335 4871 manager.go:324] Recovery completed Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.949384 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.951766 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.951872 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.951925 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.953282 4871 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.953311 4871 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.953339 4871 state_mem.go:36] "Initialized new in-memory state store" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.971468 4871 policy_none.go:49] "None policy: Start" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.975078 4871 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.975126 4871 state_mem.go:35] "Initializing new in-memory state store" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.977573 4871 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.980890 4871 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.980959 4871 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 07 22:08:46 crc kubenswrapper[4871]: I1007 22:08:46.981010 4871 kubelet.go:2335] "Starting kubelet main sync loop" Oct 07 22:08:46 crc kubenswrapper[4871]: E1007 22:08:46.981090 4871 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 07 22:08:46 crc kubenswrapper[4871]: W1007 22:08:46.982892 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:46 crc kubenswrapper[4871]: E1007 22:08:46.983000 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:46 crc kubenswrapper[4871]: E1007 22:08:46.983127 4871 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.054599 4871 manager.go:334] "Starting Device Plugin manager" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.054699 4871 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.054726 4871 server.go:79] "Starting device plugin registration server" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.055390 4871 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.055578 4871 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.055775 4871 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.055906 4871 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.055921 4871 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 07 22:08:47 crc kubenswrapper[4871]: E1007 22:08:47.067323 4871 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.081614 4871 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.081788 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.083482 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.083588 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.083607 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.083851 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.085063 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.085183 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.085287 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.085321 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.085333 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.085523 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.085958 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.086022 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.086878 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.086895 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.086908 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.086921 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.086951 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.086966 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.087179 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.087183 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.087212 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.087219 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.087246 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: E1007 22:08:47.087365 4871 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="400ms" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.087588 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.088078 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.088122 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.088140 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.088764 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.088855 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.088874 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.089101 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.089201 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.089232 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.090302 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.090330 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.090342 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.090445 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.090527 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.090560 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.090752 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.090851 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.092420 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.092480 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.092503 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.130021 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.130108 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.158416 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.160432 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.160487 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.160505 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.160546 4871 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 22:08:47 crc kubenswrapper[4871]: E1007 22:08:47.161301 4871 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232087 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232170 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232221 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232276 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232377 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232419 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232451 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232482 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232591 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232674 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232716 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232749 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232820 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232859 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232890 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.232954 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.233136 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334526 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334634 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334687 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334733 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334754 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334777 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334865 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334884 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334948 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334985 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334928 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334961 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335049 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335090 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335160 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335173 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.334905 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335208 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335231 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335254 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335280 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335273 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335332 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335345 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335404 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.335467 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.362291 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.364635 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.364698 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.364716 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.364756 4871 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 22:08:47 crc kubenswrapper[4871]: E1007 22:08:47.365492 4871 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.438172 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.448863 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.477534 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: E1007 22:08:47.490005 4871 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="800ms" Oct 07 22:08:47 crc kubenswrapper[4871]: W1007 22:08:47.497381 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-150abae549e7fbc603c40b8300082548067e97ee78d2ee9179c0a7041d84a35d WatchSource:0}: Error finding container 150abae549e7fbc603c40b8300082548067e97ee78d2ee9179c0a7041d84a35d: Status 404 returned error can't find the container with id 150abae549e7fbc603c40b8300082548067e97ee78d2ee9179c0a7041d84a35d Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.499001 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: W1007 22:08:47.501850 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-21eed7b396e30f96f852da914a2765305599b2b952b6c1a9654a4005d63b7735 WatchSource:0}: Error finding container 21eed7b396e30f96f852da914a2765305599b2b952b6c1a9654a4005d63b7735: Status 404 returned error can't find the container with id 21eed7b396e30f96f852da914a2765305599b2b952b6c1a9654a4005d63b7735 Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.510639 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:47 crc kubenswrapper[4871]: W1007 22:08:47.519635 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-67f9114be9b5f366b0357676ada85c2a64f6f5880e4c3a7dcbe6c2803ac31b3f WatchSource:0}: Error finding container 67f9114be9b5f366b0357676ada85c2a64f6f5880e4c3a7dcbe6c2803ac31b3f: Status 404 returned error can't find the container with id 67f9114be9b5f366b0357676ada85c2a64f6f5880e4c3a7dcbe6c2803ac31b3f Oct 07 22:08:47 crc kubenswrapper[4871]: W1007 22:08:47.522918 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-57863af6c31aecfc9d6c12dbc792720ef907b4ec85af72fbfd66d48e3e14ab3e WatchSource:0}: Error finding container 57863af6c31aecfc9d6c12dbc792720ef907b4ec85af72fbfd66d48e3e14ab3e: Status 404 returned error can't find the container with id 57863af6c31aecfc9d6c12dbc792720ef907b4ec85af72fbfd66d48e3e14ab3e Oct 07 22:08:47 crc kubenswrapper[4871]: W1007 22:08:47.533847 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-40d0fa97d2f065449edc9db795c235bf079493077fa04c619497709276b182fc WatchSource:0}: Error finding container 40d0fa97d2f065449edc9db795c235bf079493077fa04c619497709276b182fc: Status 404 returned error can't find the container with id 40d0fa97d2f065449edc9db795c235bf079493077fa04c619497709276b182fc Oct 07 22:08:47 crc kubenswrapper[4871]: W1007 22:08:47.753452 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:47 crc kubenswrapper[4871]: E1007 22:08:47.753596 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.766052 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.767667 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.767759 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.767784 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.767874 4871 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 22:08:47 crc kubenswrapper[4871]: E1007 22:08:47.768635 4871 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 07 22:08:47 crc kubenswrapper[4871]: W1007 22:08:47.866054 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:47 crc kubenswrapper[4871]: E1007 22:08:47.866195 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.882496 4871 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.987946 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"40d0fa97d2f065449edc9db795c235bf079493077fa04c619497709276b182fc"} Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.989476 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"57863af6c31aecfc9d6c12dbc792720ef907b4ec85af72fbfd66d48e3e14ab3e"} Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.992775 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"67f9114be9b5f366b0357676ada85c2a64f6f5880e4c3a7dcbe6c2803ac31b3f"} Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.994630 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"150abae549e7fbc603c40b8300082548067e97ee78d2ee9179c0a7041d84a35d"} Oct 07 22:08:47 crc kubenswrapper[4871]: I1007 22:08:47.996734 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"21eed7b396e30f96f852da914a2765305599b2b952b6c1a9654a4005d63b7735"} Oct 07 22:08:48 crc kubenswrapper[4871]: W1007 22:08:48.173158 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:48 crc kubenswrapper[4871]: E1007 22:08:48.173569 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:48 crc kubenswrapper[4871]: E1007 22:08:48.291640 4871 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="1.6s" Oct 07 22:08:48 crc kubenswrapper[4871]: W1007 22:08:48.324398 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:48 crc kubenswrapper[4871]: E1007 22:08:48.324548 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:48 crc kubenswrapper[4871]: I1007 22:08:48.569608 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:48 crc kubenswrapper[4871]: I1007 22:08:48.572610 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:48 crc kubenswrapper[4871]: I1007 22:08:48.572677 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:48 crc kubenswrapper[4871]: I1007 22:08:48.572698 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:48 crc kubenswrapper[4871]: I1007 22:08:48.572738 4871 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 22:08:48 crc kubenswrapper[4871]: E1007 22:08:48.573496 4871 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 07 22:08:48 crc kubenswrapper[4871]: I1007 22:08:48.881988 4871 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.002959 4871 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be" exitCode=0 Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.003079 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be"} Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.003883 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.005374 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.005412 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.005426 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.006122 4871 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d71d52a589fcf8738c5703836afc84c8d6c39ce59f971186f68f492ca55d0c24" exitCode=0 Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.006228 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d71d52a589fcf8738c5703836afc84c8d6c39ce59f971186f68f492ca55d0c24"} Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.006270 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.007897 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.008120 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.008240 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.008264 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.009575 4871 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d" exitCode=0 Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.009693 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d"} Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.009872 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.009903 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.009914 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.010358 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.013354 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.013390 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.013403 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.013759 4871 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4" exitCode=0 Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.013944 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.013945 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4"} Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.014782 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.014819 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.014828 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.018778 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233"} Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.018832 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7"} Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.018842 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a"} Oct 07 22:08:49 crc kubenswrapper[4871]: I1007 22:08:49.881353 4871 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:49 crc kubenswrapper[4871]: E1007 22:08:49.893023 4871 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="3.2s" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.025501 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f"} Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.025785 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.027597 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.027657 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.027680 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.033743 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459"} Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.033852 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba"} Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.033882 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947"} Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.033905 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0"} Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.036526 4871 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e69795d8b7f8c5469f961ad31b77d1621a957248111c29429e6cb9b7fd951712" exitCode=0 Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.036628 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e69795d8b7f8c5469f961ad31b77d1621a957248111c29429e6cb9b7fd951712"} Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.036911 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.038371 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.038408 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.038424 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.040965 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.040962 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e62ddf0800fa21c83bb765ca4ae13168e4ba543ae0b9ae2c52a264b019b17561"} Oct 07 22:08:50 crc kubenswrapper[4871]: W1007 22:08:50.043155 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:50 crc kubenswrapper[4871]: E1007 22:08:50.043240 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.043399 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.043416 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.043426 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.047844 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d"} Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.048010 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5"} Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.048034 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84"} Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.048099 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.050153 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.050182 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.050196 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.175538 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.177541 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.177600 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.177614 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:50 crc kubenswrapper[4871]: I1007 22:08:50.177648 4871 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 22:08:50 crc kubenswrapper[4871]: E1007 22:08:50.178236 4871 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 07 22:08:50 crc kubenswrapper[4871]: W1007 22:08:50.204536 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:50 crc kubenswrapper[4871]: E1007 22:08:50.204643 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:50 crc kubenswrapper[4871]: W1007 22:08:50.395604 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 07 22:08:50 crc kubenswrapper[4871]: E1007 22:08:50.395742 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.056686 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3e6e16a70dab662817d0acd3fe867fa75516c9b0f7a24487137095847ea85818"} Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.056855 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.058420 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.058455 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.058465 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.060933 4871 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c682fca106747948b2a67653392b9fa3a52de37fba7b46879c31d8715550eb1f" exitCode=0 Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.061128 4871 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.061197 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.062215 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c682fca106747948b2a67653392b9fa3a52de37fba7b46879c31d8715550eb1f"} Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.062262 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.062323 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.062889 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.064654 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.064674 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.064682 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.065381 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.065404 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.065414 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.065470 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.065505 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.065520 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.065668 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.065742 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.065849 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:51 crc kubenswrapper[4871]: I1007 22:08:51.602976 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.037387 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.069093 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f69641b781c01f9b2eabebe2fdd75d63912f188df46ebd9dc7f0752494e7ef33"} Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.069159 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b669166d30c9f0907fda36f4d5cebcd4355a8956b9df958faea1cae5729898f1"} Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.069178 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ff0f385373614a69f1bd118663ca6141796d936f4aa21fb12ab1aed05bf9d1dc"} Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.069239 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.069281 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.070676 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.070719 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.070736 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.070708 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.070929 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:52 crc kubenswrapper[4871]: I1007 22:08:52.070947 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.078429 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"200734ed77ea96f52f1dc5338444a9139d6067282f678293ab55290ae684ed04"} Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.078495 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.078496 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.078504 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3fb74c55e926664f479d24d701972388b183d7c203e3b66ca253a3669c1a69c8"} Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.080289 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.080346 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.080364 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.080307 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.080509 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.080537 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.378906 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.380895 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.380998 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.381036 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:53 crc kubenswrapper[4871]: I1007 22:08:53.381099 4871 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.081178 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.082600 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.082669 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.082691 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.345010 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.345259 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.347116 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.347169 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.347192 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.434282 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.434658 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.436716 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.436787 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:54 crc kubenswrapper[4871]: I1007 22:08:54.436852 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.144321 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.144600 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.146397 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.146471 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.146493 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.661343 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.661636 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.663472 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.663525 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.663544 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.672087 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.672233 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.673661 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.673706 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.673725 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:55 crc kubenswrapper[4871]: I1007 22:08:55.683544 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:56 crc kubenswrapper[4871]: I1007 22:08:56.087081 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:56 crc kubenswrapper[4871]: I1007 22:08:56.088484 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:56 crc kubenswrapper[4871]: I1007 22:08:56.088556 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:56 crc kubenswrapper[4871]: I1007 22:08:56.088578 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:56 crc kubenswrapper[4871]: I1007 22:08:56.261091 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:57 crc kubenswrapper[4871]: E1007 22:08:57.067535 4871 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 07 22:08:57 crc kubenswrapper[4871]: I1007 22:08:57.089652 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:57 crc kubenswrapper[4871]: I1007 22:08:57.090832 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:57 crc kubenswrapper[4871]: I1007 22:08:57.090881 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:57 crc kubenswrapper[4871]: I1007 22:08:57.090896 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:59 crc kubenswrapper[4871]: I1007 22:08:59.227303 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:59 crc kubenswrapper[4871]: I1007 22:08:59.227545 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:08:59 crc kubenswrapper[4871]: I1007 22:08:59.229418 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:08:59 crc kubenswrapper[4871]: I1007 22:08:59.229473 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:08:59 crc kubenswrapper[4871]: I1007 22:08:59.229491 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:08:59 crc kubenswrapper[4871]: I1007 22:08:59.234372 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:08:59 crc kubenswrapper[4871]: I1007 22:08:59.261681 4871 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 22:08:59 crc kubenswrapper[4871]: I1007 22:08:59.261869 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 22:09:00 crc kubenswrapper[4871]: I1007 22:09:00.099303 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:00 crc kubenswrapper[4871]: I1007 22:09:00.104567 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:00 crc kubenswrapper[4871]: I1007 22:09:00.104707 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:00 crc kubenswrapper[4871]: I1007 22:09:00.104746 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:00 crc kubenswrapper[4871]: W1007 22:09:00.756466 4871 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 07 22:09:00 crc kubenswrapper[4871]: I1007 22:09:00.756613 4871 trace.go:236] Trace[1415843412]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 22:08:50.754) (total time: 10001ms): Oct 07 22:09:00 crc kubenswrapper[4871]: Trace[1415843412]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (22:09:00.756) Oct 07 22:09:00 crc kubenswrapper[4871]: Trace[1415843412]: [10.001847016s] [10.001847016s] END Oct 07 22:09:00 crc kubenswrapper[4871]: E1007 22:09:00.756642 4871 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 07 22:09:00 crc kubenswrapper[4871]: I1007 22:09:00.883395 4871 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.104376 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.107211 4871 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3e6e16a70dab662817d0acd3fe867fa75516c9b0f7a24487137095847ea85818" exitCode=255 Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.107299 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3e6e16a70dab662817d0acd3fe867fa75516c9b0f7a24487137095847ea85818"} Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.107684 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.109174 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.109237 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.109258 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.110327 4871 scope.go:117] "RemoveContainer" containerID="3e6e16a70dab662817d0acd3fe867fa75516c9b0f7a24487137095847ea85818" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.460727 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.461063 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.462682 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.462737 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.462750 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:01 crc kubenswrapper[4871]: I1007 22:09:01.505768 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.092871 4871 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.092972 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.098207 4871 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.098288 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.113241 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.115737 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23"} Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.115856 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.116013 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.116764 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.116813 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.116823 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.117283 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.117313 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.117326 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:02 crc kubenswrapper[4871]: I1007 22:09:02.129974 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 07 22:09:03 crc kubenswrapper[4871]: I1007 22:09:03.117529 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:03 crc kubenswrapper[4871]: I1007 22:09:03.118568 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:03 crc kubenswrapper[4871]: I1007 22:09:03.118626 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:03 crc kubenswrapper[4871]: I1007 22:09:03.118640 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:04 crc kubenswrapper[4871]: I1007 22:09:04.444394 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:09:04 crc kubenswrapper[4871]: I1007 22:09:04.444624 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:04 crc kubenswrapper[4871]: I1007 22:09:04.444883 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:09:04 crc kubenswrapper[4871]: I1007 22:09:04.446377 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:04 crc kubenswrapper[4871]: I1007 22:09:04.446458 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:04 crc kubenswrapper[4871]: I1007 22:09:04.446485 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:04 crc kubenswrapper[4871]: I1007 22:09:04.452430 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:09:05 crc kubenswrapper[4871]: I1007 22:09:05.124341 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:05 crc kubenswrapper[4871]: I1007 22:09:05.125711 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:05 crc kubenswrapper[4871]: I1007 22:09:05.125772 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:05 crc kubenswrapper[4871]: I1007 22:09:05.125823 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:06 crc kubenswrapper[4871]: I1007 22:09:06.119341 4871 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 07 22:09:06 crc kubenswrapper[4871]: I1007 22:09:06.127019 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:06 crc kubenswrapper[4871]: I1007 22:09:06.128598 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:06 crc kubenswrapper[4871]: I1007 22:09:06.128683 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:06 crc kubenswrapper[4871]: I1007 22:09:06.128702 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.067968 4871 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.082503 4871 trace.go:236] Trace[1659663093]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 22:08:54.945) (total time: 12136ms): Oct 07 22:09:07 crc kubenswrapper[4871]: Trace[1659663093]: ---"Objects listed" error: 12136ms (22:09:07.082) Oct 07 22:09:07 crc kubenswrapper[4871]: Trace[1659663093]: [12.136531295s] [12.136531295s] END Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.082956 4871 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.083179 4871 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.084274 4871 trace.go:236] Trace[945772561]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 22:08:53.856) (total time: 13228ms): Oct 07 22:09:07 crc kubenswrapper[4871]: Trace[945772561]: ---"Objects listed" error: 13228ms (22:09:07.084) Oct 07 22:09:07 crc kubenswrapper[4871]: Trace[945772561]: [13.228080543s] [13.228080543s] END Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.084296 4871 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.084397 4871 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.084415 4871 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.084780 4871 trace.go:236] Trace[1448149412]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 22:08:56.469) (total time: 10614ms): Oct 07 22:09:07 crc kubenswrapper[4871]: Trace[1448149412]: ---"Objects listed" error: 10614ms (22:09:07.084) Oct 07 22:09:07 crc kubenswrapper[4871]: Trace[1448149412]: [10.614393394s] [10.614393394s] END Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.084963 4871 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.125090 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.132876 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.876267 4871 apiserver.go:52] "Watching apiserver" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.878997 4871 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.879365 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.879938 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.879964 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.880117 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.880280 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.880378 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.880408 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.880427 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.880701 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.880761 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.883149 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.883486 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.883548 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.883625 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.883684 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.883760 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.883776 4871 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.883706 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.883743 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.886465 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889156 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889188 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889207 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889224 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889241 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889261 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889279 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889294 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889310 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889331 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889346 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889364 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889406 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889424 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889443 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889460 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889496 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889532 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889549 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889565 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889607 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889627 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889651 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889690 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889681 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889706 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889861 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889944 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.889985 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890022 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890056 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890092 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890127 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890163 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890196 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890232 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890324 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890378 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890416 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890453 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890492 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890530 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890564 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890598 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890639 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890677 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890710 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890744 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890777 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890849 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890887 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890924 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890958 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890994 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891026 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891058 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891113 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891151 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891182 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891217 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891249 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891282 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891317 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891349 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891380 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891411 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891445 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891478 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891512 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891543 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891580 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891611 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891642 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891672 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891705 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891739 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891773 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891877 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891911 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891946 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891980 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892014 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892048 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892079 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892111 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892143 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892197 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892231 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892268 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892300 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892333 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892368 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892403 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892435 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892466 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892499 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892541 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892577 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892616 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892650 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892684 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892721 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892753 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892786 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893013 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893075 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893112 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893146 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893184 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893219 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893258 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893300 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893343 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893381 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893415 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893451 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893498 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893535 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893576 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893609 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893646 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893684 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893726 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893763 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893822 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893877 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893934 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893968 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894005 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894045 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894079 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894118 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894166 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894202 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894242 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894275 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894308 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894344 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894438 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894481 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894516 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894550 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894586 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894621 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894657 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894694 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894736 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894775 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894838 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894875 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895008 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895053 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895094 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895129 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895164 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895201 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895240 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890060 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895282 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890292 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.890516 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891054 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891685 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891764 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891917 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.891994 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895427 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892231 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892323 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892462 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.892512 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893042 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893147 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893411 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893643 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893733 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.893928 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894368 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894829 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.894911 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895182 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895272 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895276 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895906 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.896105 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.896226 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.895324 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.896704 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.896749 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.896783 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.896920 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.896981 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897017 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897054 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897098 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897131 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897165 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897272 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897309 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897340 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897372 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897399 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897428 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897630 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897659 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897684 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897713 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897741 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897769 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897843 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897873 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897921 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897950 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.897978 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898006 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898032 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898057 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898082 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898123 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898151 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898181 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898213 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898242 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898267 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898333 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898370 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898398 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898435 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898465 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898493 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898525 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898552 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898578 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898602 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898631 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898658 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898683 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898724 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898827 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898848 4871 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898860 4871 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898873 4871 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898883 4871 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898893 4871 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898942 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.898998 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899023 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899036 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899048 4871 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899059 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899069 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899080 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899091 4871 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899102 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899112 4871 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899123 4871 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899134 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899145 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899155 4871 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899167 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899178 4871 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899189 4871 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899199 4871 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899250 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899276 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899286 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.899295 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.896619 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.896812 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.899394 4871 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.903462 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.903902 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.904734 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.904838 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.904982 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.905197 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.904872 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.905575 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.905673 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.905852 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.906115 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.906233 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.906354 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.907651 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.907742 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.907952 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.907992 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.908317 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.908400 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.908596 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.908607 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.908671 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.908925 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.909247 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.909305 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.909527 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.909616 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.909838 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.910083 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.910131 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.910400 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.910497 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.904172 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.910867 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.911623 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:08.411590548 +0000 UTC m=+22.214288621 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.912003 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.915114 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.912009 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.915135 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.912166 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.912580 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.912733 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.912893 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.913883 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.913894 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.913914 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.913932 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.914001 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.914336 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.914336 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.914477 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.914770 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.914780 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.915099 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.912255 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.915974 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.916029 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:09:08.41506519 +0000 UTC m=+22.217763273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.916243 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.916299 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.918194 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.918352 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.916346 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.920905 4871 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.922026 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.922493 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.922956 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.923235 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.923142 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.923453 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.923942 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.924967 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.925439 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.926052 4871 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.926204 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:08.426178225 +0000 UTC m=+22.228876298 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.937298 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.937937 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.938074 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.938207 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.938912 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.941050 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.941470 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.941718 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.942238 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.942464 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.942661 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.942656 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.942723 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.942880 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.942880 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.943088 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.943115 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.943131 4871 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.943195 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.943212 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:08.443186507 +0000 UTC m=+22.245884570 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.943275 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.943505 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.943529 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.944032 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.944159 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.944171 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.944289 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.944417 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.944453 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.944636 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.944635 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.944865 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.944886 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.944882 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.944896 4871 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:07 crc kubenswrapper[4871]: E1007 22:09:07.945098 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:08.445067047 +0000 UTC m=+22.247765120 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.946033 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.947223 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.947423 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.947596 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.948339 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.948541 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.949185 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.950379 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.950484 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.950660 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.950955 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.951191 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.951413 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.951454 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.951480 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.951489 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.951740 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.952143 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.952567 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.953955 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.954007 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.954269 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.954512 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.955105 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.955486 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.956004 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.956336 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.956438 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.956715 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.958324 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.958604 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.958631 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.959420 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.959714 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.960563 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.961449 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.961930 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.961968 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.962007 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.962244 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.962116 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.962277 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.962530 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.962930 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.962905 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.963402 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.964388 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.964877 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.970409 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.971409 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.971784 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.972201 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.972440 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.972495 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.972536 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.972651 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.973304 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.973424 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.974542 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.976203 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.976306 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.979004 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.979097 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.979500 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.979713 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.980876 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.981999 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.982917 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.983190 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.983901 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.988132 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.989049 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.988200 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.988548 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.989260 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:09:07 crc kubenswrapper[4871]: W1007 22:09:07.989261 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-cb9933287c3d1ef779ec5e5e501a75f07dcab1f3a1dc1444313000934e4f70f7 WatchSource:0}: Error finding container cb9933287c3d1ef779ec5e5e501a75f07dcab1f3a1dc1444313000934e4f70f7: Status 404 returned error can't find the container with id cb9933287c3d1ef779ec5e5e501a75f07dcab1f3a1dc1444313000934e4f70f7 Oct 07 22:09:07 crc kubenswrapper[4871]: I1007 22:09:07.994065 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000042 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000208 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000269 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000413 4871 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000436 4871 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000450 4871 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000497 4871 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000509 4871 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000522 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000533 4871 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000624 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000641 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000653 4871 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000704 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000718 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000730 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000743 4871 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000863 4871 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000876 4871 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000891 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000955 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.000969 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001035 4871 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001052 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001104 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001121 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001134 4871 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001148 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001148 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001174 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001193 4871 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001206 4871 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001225 4871 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001237 4871 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001251 4871 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001264 4871 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001278 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001290 4871 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001304 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001315 4871 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001327 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001339 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001351 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001364 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001376 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001434 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001447 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001488 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001500 4871 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001514 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001532 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001548 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001561 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001575 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001590 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001602 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001616 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001627 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001628 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001688 4871 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001706 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001719 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001730 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001743 4871 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001756 4871 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001768 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001781 4871 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001823 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001836 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001848 4871 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001861 4871 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001873 4871 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001886 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001900 4871 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001912 4871 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001927 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001939 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001951 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001968 4871 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001980 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.001993 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002004 4871 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002019 4871 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002033 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002044 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002056 4871 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002068 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002078 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002092 4871 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002104 4871 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002115 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002127 4871 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002139 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002154 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002168 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002180 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002193 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002207 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002218 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002230 4871 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002242 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002254 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002265 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002277 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002289 4871 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002301 4871 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002312 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002324 4871 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002336 4871 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002348 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002360 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002371 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002382 4871 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002394 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002413 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002427 4871 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002441 4871 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002455 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002471 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002483 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002495 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002508 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002518 4871 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002551 4871 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002563 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002574 4871 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002586 4871 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002596 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002609 4871 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002621 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002633 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002646 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002658 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002669 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002680 4871 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002692 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002704 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002716 4871 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002727 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002739 4871 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002751 4871 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002763 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002775 4871 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002803 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002816 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002828 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002839 4871 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002850 4871 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002861 4871 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002878 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002890 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002901 4871 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002912 4871 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002925 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002936 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002948 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002958 4871 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002970 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002982 4871 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.002994 4871 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.003008 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.003020 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.003030 4871 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.003041 4871 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.003052 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.008995 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.009019 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.016738 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.017183 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.023227 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.026672 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.104233 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.104716 4871 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.104730 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.104742 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.104754 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.134019 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cb9933287c3d1ef779ec5e5e501a75f07dcab1f3a1dc1444313000934e4f70f7"} Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.136037 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.136596 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.138845 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23"} Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.138852 4871 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23" exitCode=255 Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.138948 4871 scope.go:117] "RemoveContainer" containerID="3e6e16a70dab662817d0acd3fe867fa75516c9b0f7a24487137095847ea85818" Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.148127 4871 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.165613 4871 scope.go:117] "RemoveContainer" containerID="86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23" Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.165962 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.169454 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.171937 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.199215 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.224119 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.225366 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.244153 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: W1007 22:09:08.245289 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-29a059c3d9e622300222601c314d283e791b2033731ca1d81d56b489e57f8c27 WatchSource:0}: Error finding container 29a059c3d9e622300222601c314d283e791b2033731ca1d81d56b489e57f8c27: Status 404 returned error can't find the container with id 29a059c3d9e622300222601c314d283e791b2033731ca1d81d56b489e57f8c27 Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.265234 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.274763 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.287535 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.299223 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.511056 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.511135 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.511164 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.511186 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.511206 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511337 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511358 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511371 4871 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511424 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:09.511407727 +0000 UTC m=+23.314105800 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511475 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:09:09.511469988 +0000 UTC m=+23.314168061 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511514 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511524 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511531 4871 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511552 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:09.51154577 +0000 UTC m=+23.314243843 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511595 4871 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511615 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:09.511608592 +0000 UTC m=+23.314306665 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511642 4871 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:08 crc kubenswrapper[4871]: E1007 22:09:08.511661 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:09.511655643 +0000 UTC m=+23.314353716 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.753502 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-vl6pr"] Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.753919 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-hp4jl"] Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.754054 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vl6pr" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.754093 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-nv4jc"] Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.754354 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.754574 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.758567 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.759164 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.759670 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.759906 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.760098 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.760318 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.760455 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.759744 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-h66dd"] Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.759823 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.761434 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.761501 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.761640 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.761631 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.761900 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.766909 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.769244 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.769774 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.789825 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e6e16a70dab662817d0acd3fe867fa75516c9b0f7a24487137095847ea85818\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:00Z\\\",\\\"message\\\":\\\"W1007 22:08:50.319717 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 22:08:50.320061 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759874930 cert, and key in /tmp/serving-cert-1009882592/serving-signer.crt, /tmp/serving-cert-1009882592/serving-signer.key\\\\nI1007 22:08:50.460623 1 observer_polling.go:159] Starting file observer\\\\nW1007 22:08:50.463779 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 22:08:50.463991 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:08:50.465912 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1009882592/tls.crt::/tmp/serving-cert-1009882592/tls.key\\\\\\\"\\\\nF1007 22:09:00.798356 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.810917 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813458 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-run-k8s-cni-cncf-io\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813492 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2q9t\" (UniqueName: \"kubernetes.io/projected/53708429-59b4-4319-bdb7-8a922a551e59-kube-api-access-c2q9t\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813604 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-cnibin\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813655 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-run-multus-certs\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813720 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-var-lib-kubelet\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813760 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/817bca9c-66e5-440f-a773-d49006702b3a-hosts-file\") pod \"node-resolver-vl6pr\" (UID: \"817bca9c-66e5-440f-a773-d49006702b3a\") " pod="openshift-dns/node-resolver-vl6pr" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813808 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-rootfs\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813827 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-multus-conf-dir\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813847 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-cnibin\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813864 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813884 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-etc-kubernetes\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813911 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-system-cni-dir\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813928 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-var-lib-cni-bin\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813945 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-os-release\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813965 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ad179ea2-56e8-4e5e-a72c-599c797918d1-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.813986 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-system-cni-dir\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814017 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-multus-socket-dir-parent\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814047 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsnsj\" (UniqueName: \"kubernetes.io/projected/817bca9c-66e5-440f-a773-d49006702b3a-kube-api-access-vsnsj\") pod \"node-resolver-vl6pr\" (UID: \"817bca9c-66e5-440f-a773-d49006702b3a\") " pod="openshift-dns/node-resolver-vl6pr" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814088 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-mcd-auth-proxy-config\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814116 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8x84\" (UniqueName: \"kubernetes.io/projected/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-kube-api-access-m8x84\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814143 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/53708429-59b4-4319-bdb7-8a922a551e59-cni-binary-copy\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814174 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-run-netns\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814203 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-var-lib-cni-multus\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814238 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-multus-cni-dir\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814261 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-os-release\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814285 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ad179ea2-56e8-4e5e-a72c-599c797918d1-cni-binary-copy\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814311 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shmsz\" (UniqueName: \"kubernetes.io/projected/ad179ea2-56e8-4e5e-a72c-599c797918d1-kube-api-access-shmsz\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814381 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-proxy-tls\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814409 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-hostroot\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.814514 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/53708429-59b4-4319-bdb7-8a922a551e59-multus-daemon-config\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.823470 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.836603 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.849224 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.865388 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.881568 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.893510 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.904749 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915561 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-system-cni-dir\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915613 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-multus-socket-dir-parent\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915636 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsnsj\" (UniqueName: \"kubernetes.io/projected/817bca9c-66e5-440f-a773-d49006702b3a-kube-api-access-vsnsj\") pod \"node-resolver-vl6pr\" (UID: \"817bca9c-66e5-440f-a773-d49006702b3a\") " pod="openshift-dns/node-resolver-vl6pr" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915660 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-mcd-auth-proxy-config\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915680 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8x84\" (UniqueName: \"kubernetes.io/projected/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-kube-api-access-m8x84\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915702 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/53708429-59b4-4319-bdb7-8a922a551e59-cni-binary-copy\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915722 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-run-netns\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915746 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-var-lib-cni-multus\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915741 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-system-cni-dir\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915769 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-multus-cni-dir\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915859 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-os-release\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915881 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ad179ea2-56e8-4e5e-a72c-599c797918d1-cni-binary-copy\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915900 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shmsz\" (UniqueName: \"kubernetes.io/projected/ad179ea2-56e8-4e5e-a72c-599c797918d1-kube-api-access-shmsz\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915931 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-proxy-tls\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915946 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-hostroot\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915960 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-multus-cni-dir\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.915964 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-multus-socket-dir-parent\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916007 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-hostroot\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916017 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/53708429-59b4-4319-bdb7-8a922a551e59-multus-daemon-config\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916081 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-run-k8s-cni-cncf-io\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916109 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2q9t\" (UniqueName: \"kubernetes.io/projected/53708429-59b4-4319-bdb7-8a922a551e59-kube-api-access-c2q9t\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916135 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-cnibin\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916161 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-run-multus-certs\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916199 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-var-lib-kubelet\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916226 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/817bca9c-66e5-440f-a773-d49006702b3a-hosts-file\") pod \"node-resolver-vl6pr\" (UID: \"817bca9c-66e5-440f-a773-d49006702b3a\") " pod="openshift-dns/node-resolver-vl6pr" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916282 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-rootfs\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916304 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-multus-conf-dir\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916328 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-cnibin\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916350 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916374 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-etc-kubernetes\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916421 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-system-cni-dir\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916449 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-var-lib-cni-bin\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916473 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-os-release\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916497 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ad179ea2-56e8-4e5e-a72c-599c797918d1-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916856 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ad179ea2-56e8-4e5e-a72c-599c797918d1-cni-binary-copy\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.916933 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/53708429-59b4-4319-bdb7-8a922a551e59-multus-daemon-config\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917038 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-mcd-auth-proxy-config\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917234 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-var-lib-cni-bin\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917294 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-system-cni-dir\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917305 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ad179ea2-56e8-4e5e-a72c-599c797918d1-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917322 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-run-k8s-cni-cncf-io\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917457 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-os-release\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917502 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/817bca9c-66e5-440f-a773-d49006702b3a-hosts-file\") pod \"node-resolver-vl6pr\" (UID: \"817bca9c-66e5-440f-a773-d49006702b3a\") " pod="openshift-dns/node-resolver-vl6pr" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917535 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-cnibin\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917582 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-run-multus-certs\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917586 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-etc-kubernetes\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917600 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-var-lib-kubelet\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917612 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-multus-conf-dir\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917637 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-cnibin\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917666 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-run-netns\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917693 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-host-var-lib-cni-multus\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917722 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-rootfs\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917806 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/53708429-59b4-4319-bdb7-8a922a551e59-cni-binary-copy\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917843 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ad179ea2-56e8-4e5e-a72c-599c797918d1-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.917863 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/53708429-59b4-4319-bdb7-8a922a551e59-os-release\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.923189 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.954148 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.982103 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-proxy-tls\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.982194 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8x84\" (UniqueName: \"kubernetes.io/projected/ef3b09cf-b090-4f2c-892d-ab7f7aee3129-kube-api-access-m8x84\") pod \"machine-config-daemon-hp4jl\" (UID: \"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\") " pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.983306 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shmsz\" (UniqueName: \"kubernetes.io/projected/ad179ea2-56e8-4e5e-a72c-599c797918d1-kube-api-access-shmsz\") pod \"multus-additional-cni-plugins-h66dd\" (UID: \"ad179ea2-56e8-4e5e-a72c-599c797918d1\") " pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.983397 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsnsj\" (UniqueName: \"kubernetes.io/projected/817bca9c-66e5-440f-a773-d49006702b3a-kube-api-access-vsnsj\") pod \"node-resolver-vl6pr\" (UID: \"817bca9c-66e5-440f-a773-d49006702b3a\") " pod="openshift-dns/node-resolver-vl6pr" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.985448 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.986163 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.987502 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.988148 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.989322 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.989994 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.990142 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2q9t\" (UniqueName: \"kubernetes.io/projected/53708429-59b4-4319-bdb7-8a922a551e59-kube-api-access-c2q9t\") pod \"multus-nv4jc\" (UID: \"53708429-59b4-4319-bdb7-8a922a551e59\") " pod="openshift-multus/multus-nv4jc" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.990642 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.991683 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.992530 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.994229 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.994470 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.996125 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.997085 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.998227 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.998974 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 07 22:09:08 crc kubenswrapper[4871]: I1007 22:09:08.999512 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.002374 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.003154 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.004230 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.005012 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.005773 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.006880 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.007619 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.008263 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.009614 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.010129 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.011392 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.012377 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.013397 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.014125 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.015200 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.015780 4871 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.016008 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.018541 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.018666 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e6e16a70dab662817d0acd3fe867fa75516c9b0f7a24487137095847ea85818\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:00Z\\\",\\\"message\\\":\\\"W1007 22:08:50.319717 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 22:08:50.320061 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759874930 cert, and key in /tmp/serving-cert-1009882592/serving-signer.crt, /tmp/serving-cert-1009882592/serving-signer.key\\\\nI1007 22:08:50.460623 1 observer_polling.go:159] Starting file observer\\\\nW1007 22:08:50.463779 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 22:08:50.463991 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:08:50.465912 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1009882592/tls.crt::/tmp/serving-cert-1009882592/tls.key\\\\\\\"\\\\nF1007 22:09:00.798356 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.019153 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.019562 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.021455 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.022635 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.023272 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.024377 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.025078 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.025928 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.026527 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.027590 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.028682 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.029312 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.029885 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.030757 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.031651 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.032679 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.033262 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.033265 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.034293 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.034958 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.035760 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.036735 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.044918 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.057414 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.070697 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.074777 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vl6pr" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.082132 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.084883 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.089132 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nv4jc" Oct 07 22:09:09 crc kubenswrapper[4871]: W1007 22:09:09.089567 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod817bca9c_66e5_440f_a773_d49006702b3a.slice/crio-0743a4a50a98df5618b3cd8be72d6b2508c47e931dff281ac6f4ce73299155f5 WatchSource:0}: Error finding container 0743a4a50a98df5618b3cd8be72d6b2508c47e931dff281ac6f4ce73299155f5: Status 404 returned error can't find the container with id 0743a4a50a98df5618b3cd8be72d6b2508c47e931dff281ac6f4ce73299155f5 Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.100613 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-h66dd" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.103650 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.116579 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: W1007 22:09:09.126944 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad179ea2_56e8_4e5e_a72c_599c797918d1.slice/crio-b02b9e1d436f81a194a7d500a74d94ae1e962716a74c4b13390ddfa94d1b3755 WatchSource:0}: Error finding container b02b9e1d436f81a194a7d500a74d94ae1e962716a74c4b13390ddfa94d1b3755: Status 404 returned error can't find the container with id b02b9e1d436f81a194a7d500a74d94ae1e962716a74c4b13390ddfa94d1b3755 Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.129974 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.132398 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-d44j5"] Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.135559 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.140513 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.141031 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.141193 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.141274 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.141344 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.141399 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.141418 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.151928 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vl6pr" event={"ID":"817bca9c-66e5-440f-a773-d49006702b3a","Type":"ContainerStarted","Data":"0743a4a50a98df5618b3cd8be72d6b2508c47e931dff281ac6f4ce73299155f5"} Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.155173 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.156727 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.158243 4871 scope.go:117] "RemoveContainer" containerID="86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23" Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.158454 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.160436 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3"} Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.160472 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46"} Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.166812 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" event={"ID":"ad179ea2-56e8-4e5e-a72c-599c797918d1","Type":"ContainerStarted","Data":"b02b9e1d436f81a194a7d500a74d94ae1e962716a74c4b13390ddfa94d1b3755"} Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.170831 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9d8cd5e6422d974c27c159133b366d8adaa16350f233430908848c43b2eddd8c"} Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.172499 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.173771 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"a2b32e2e0acbf58e5c9d855c25b06ad20ca761f8ffb6a8666ac445ebbdd7cd9f"} Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.179893 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0"} Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.179945 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"29a059c3d9e622300222601c314d283e791b2033731ca1d81d56b489e57f8c27"} Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.184507 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nv4jc" event={"ID":"53708429-59b4-4319-bdb7-8a922a551e59","Type":"ContainerStarted","Data":"0d3b90f77a4592a094559ba53382d89379b0aa6d0bf62a87276bad93e0ec7205"} Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.184629 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.198607 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.212603 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.218852 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-env-overrides\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.218882 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovn-node-metrics-cert\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.218909 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-slash\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.218943 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-ovn\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.218959 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-systemd-units\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.218977 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219004 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-bin\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219018 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-log-socket\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219046 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-etc-openvswitch\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219064 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-var-lib-openvswitch\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219084 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-config\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219100 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7hp6\" (UniqueName: \"kubernetes.io/projected/0059e7e8-7d63-4b03-81a0-b8521803f34a-kube-api-access-q7hp6\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219116 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-systemd\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219136 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-openvswitch\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219152 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-script-lib\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219182 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-node-log\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219202 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-netd\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219219 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-kubelet\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219235 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-netns\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.219265 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-ovn-kubernetes\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.224666 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.252109 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.267772 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.285218 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.308911 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.319940 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-systemd-units\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.319993 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320012 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-bin\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320027 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-log-socket\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320047 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-etc-openvswitch\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320068 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-var-lib-openvswitch\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320085 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-config\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320102 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-systemd\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320119 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7hp6\" (UniqueName: \"kubernetes.io/projected/0059e7e8-7d63-4b03-81a0-b8521803f34a-kube-api-access-q7hp6\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320140 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-openvswitch\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320156 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-script-lib\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320171 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-node-log\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320186 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-netd\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320212 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-kubelet\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320227 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-netns\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320242 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-ovn-kubernetes\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320259 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-env-overrides\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320275 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovn-node-metrics-cert\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320295 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-slash\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320312 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-ovn\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320372 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-ovn\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320440 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-openvswitch\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320488 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-systemd-units\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320525 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320559 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-bin\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320588 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-log-socket\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320618 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-etc-openvswitch\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320652 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-var-lib-openvswitch\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320766 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-netns\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320840 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-node-log\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320877 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-kubelet\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320914 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-slash\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.321009 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-script-lib\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.320988 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-netd\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.321076 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-ovn-kubernetes\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.321089 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-systemd\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.321204 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-env-overrides\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.322507 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-config\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.325135 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovn-node-metrics-cert\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.326057 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e6e16a70dab662817d0acd3fe867fa75516c9b0f7a24487137095847ea85818\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:00Z\\\",\\\"message\\\":\\\"W1007 22:08:50.319717 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 22:08:50.320061 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759874930 cert, and key in /tmp/serving-cert-1009882592/serving-signer.crt, /tmp/serving-cert-1009882592/serving-signer.key\\\\nI1007 22:08:50.460623 1 observer_polling.go:159] Starting file observer\\\\nW1007 22:08:50.463779 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 22:08:50.463991 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:08:50.465912 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1009882592/tls.crt::/tmp/serving-cert-1009882592/tls.key\\\\\\\"\\\\nF1007 22:09:00.798356 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.338820 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7hp6\" (UniqueName: \"kubernetes.io/projected/0059e7e8-7d63-4b03-81a0-b8521803f34a-kube-api-access-q7hp6\") pod \"ovnkube-node-d44j5\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.339509 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.353589 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.369772 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.385028 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.399319 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.416415 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.431181 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.445972 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.458389 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.467491 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.470966 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: W1007 22:09:09.483733 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0059e7e8_7d63_4b03_81a0_b8521803f34a.slice/crio-f24c5e6ebbe3f07e4af9c5d1a9a86bdff3fd1059ba514824a6b16f188d5c668b WatchSource:0}: Error finding container f24c5e6ebbe3f07e4af9c5d1a9a86bdff3fd1059ba514824a6b16f188d5c668b: Status 404 returned error can't find the container with id f24c5e6ebbe3f07e4af9c5d1a9a86bdff3fd1059ba514824a6b16f188d5c668b Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.489160 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.508859 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.521533 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.521705 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.521751 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.521807 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.521832 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.521998 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522025 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522040 4871 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522102 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:11.522080945 +0000 UTC m=+25.324779018 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522513 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:09:11.522500606 +0000 UTC m=+25.325198669 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522589 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522605 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522615 4871 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522647 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:11.52263815 +0000 UTC m=+25.325336223 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522692 4871 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522746 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:11.522738023 +0000 UTC m=+25.325436096 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522818 4871 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.522848 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:11.522840535 +0000 UTC m=+25.325538608 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.523633 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.555400 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.597161 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.981549 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.981616 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:09 crc kubenswrapper[4871]: I1007 22:09:09.981639 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.981710 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.981858 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:09 crc kubenswrapper[4871]: E1007 22:09:09.982122 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.191251 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nv4jc" event={"ID":"53708429-59b4-4319-bdb7-8a922a551e59","Type":"ContainerStarted","Data":"ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674"} Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.193759 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vl6pr" event={"ID":"817bca9c-66e5-440f-a773-d49006702b3a","Type":"ContainerStarted","Data":"939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7"} Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.196159 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd" exitCode=0 Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.196255 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd"} Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.196334 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"f24c5e6ebbe3f07e4af9c5d1a9a86bdff3fd1059ba514824a6b16f188d5c668b"} Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.200292 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f"} Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.200390 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002"} Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.203195 4871 generic.go:334] "Generic (PLEG): container finished" podID="ad179ea2-56e8-4e5e-a72c-599c797918d1" containerID="31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f" exitCode=0 Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.203332 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" event={"ID":"ad179ea2-56e8-4e5e-a72c-599c797918d1","Type":"ContainerDied","Data":"31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f"} Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.204566 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.225746 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.242157 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.256424 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.273417 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.302217 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.322284 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.339615 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.355534 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.368156 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.386230 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.405686 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.418327 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.430166 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.443086 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.456692 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.475873 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.496247 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.510998 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.522846 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.540267 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.555659 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.575656 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.590299 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.602773 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:10 crc kubenswrapper[4871]: I1007 22:09:10.618821 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:10Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.211434 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" event={"ID":"ad179ea2-56e8-4e5e-a72c-599c797918d1","Type":"ContainerStarted","Data":"da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde"} Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.213529 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44"} Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.222336 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.222423 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.222449 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.222469 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.222485 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.230147 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.250933 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.265751 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.284230 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.298680 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.310393 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.331495 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.347725 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.361900 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.379379 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.395171 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.412256 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.429432 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.443763 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.465843 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.480620 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.495143 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.510116 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.533099 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.540912 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.541086 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.541152 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.541200 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.541263 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541343 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:09:15.541277135 +0000 UTC m=+29.343975258 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541358 4871 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541405 4871 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541460 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541495 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541404 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541516 4871 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541557 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541509 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:15.541480881 +0000 UTC m=+29.344178974 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541589 4871 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541617 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:15.541591814 +0000 UTC m=+29.344289897 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541641 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:15.541630345 +0000 UTC m=+29.344328428 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.541660 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:15.541650185 +0000 UTC m=+29.344348268 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.560571 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.578534 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.595613 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.607270 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.622984 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.640974 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.657615 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.829234 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-xwhnv"] Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.829982 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.832908 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.833057 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.833317 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.833331 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.859870 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.876413 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.895334 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.919428 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.939775 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.944670 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk8nx\" (UniqueName: \"kubernetes.io/projected/d535a1fa-1557-4e32-8583-d8415c47c1bf-kube-api-access-jk8nx\") pod \"node-ca-xwhnv\" (UID: \"d535a1fa-1557-4e32-8583-d8415c47c1bf\") " pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.944726 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d535a1fa-1557-4e32-8583-d8415c47c1bf-serviceca\") pod \"node-ca-xwhnv\" (UID: \"d535a1fa-1557-4e32-8583-d8415c47c1bf\") " pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.944757 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d535a1fa-1557-4e32-8583-d8415c47c1bf-host\") pod \"node-ca-xwhnv\" (UID: \"d535a1fa-1557-4e32-8583-d8415c47c1bf\") " pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.957619 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.977268 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.981463 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.981454 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.981633 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.981476 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.981748 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:11 crc kubenswrapper[4871]: E1007 22:09:11.981923 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:11 crc kubenswrapper[4871]: I1007 22:09:11.992768 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:11Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.015298 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.033404 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.046195 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d535a1fa-1557-4e32-8583-d8415c47c1bf-host\") pod \"node-ca-xwhnv\" (UID: \"d535a1fa-1557-4e32-8583-d8415c47c1bf\") " pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.046324 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk8nx\" (UniqueName: \"kubernetes.io/projected/d535a1fa-1557-4e32-8583-d8415c47c1bf-kube-api-access-jk8nx\") pod \"node-ca-xwhnv\" (UID: \"d535a1fa-1557-4e32-8583-d8415c47c1bf\") " pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.046360 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d535a1fa-1557-4e32-8583-d8415c47c1bf-serviceca\") pod \"node-ca-xwhnv\" (UID: \"d535a1fa-1557-4e32-8583-d8415c47c1bf\") " pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.046378 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d535a1fa-1557-4e32-8583-d8415c47c1bf-host\") pod \"node-ca-xwhnv\" (UID: \"d535a1fa-1557-4e32-8583-d8415c47c1bf\") " pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.047553 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d535a1fa-1557-4e32-8583-d8415c47c1bf-serviceca\") pod \"node-ca-xwhnv\" (UID: \"d535a1fa-1557-4e32-8583-d8415c47c1bf\") " pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.051449 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.067700 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.072461 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk8nx\" (UniqueName: \"kubernetes.io/projected/d535a1fa-1557-4e32-8583-d8415c47c1bf-kube-api-access-jk8nx\") pod \"node-ca-xwhnv\" (UID: \"d535a1fa-1557-4e32-8583-d8415c47c1bf\") " pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.083617 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.100472 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.153244 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xwhnv" Oct 07 22:09:12 crc kubenswrapper[4871]: W1007 22:09:12.172100 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd535a1fa_1557_4e32_8583_d8415c47c1bf.slice/crio-cce5f7a908403f6695f861c6302dd01e992e0e4e30e1b5cda2e1662479e243ac WatchSource:0}: Error finding container cce5f7a908403f6695f861c6302dd01e992e0e4e30e1b5cda2e1662479e243ac: Status 404 returned error can't find the container with id cce5f7a908403f6695f861c6302dd01e992e0e4e30e1b5cda2e1662479e243ac Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.235006 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.236405 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xwhnv" event={"ID":"d535a1fa-1557-4e32-8583-d8415c47c1bf","Type":"ContainerStarted","Data":"cce5f7a908403f6695f861c6302dd01e992e0e4e30e1b5cda2e1662479e243ac"} Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.238299 4871 generic.go:334] "Generic (PLEG): container finished" podID="ad179ea2-56e8-4e5e-a72c-599c797918d1" containerID="da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde" exitCode=0 Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.238411 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" event={"ID":"ad179ea2-56e8-4e5e-a72c-599c797918d1","Type":"ContainerDied","Data":"da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde"} Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.256666 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.272264 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.297530 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.309034 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.323917 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.336041 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.352549 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.381150 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.397334 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.410520 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.426114 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.466777 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.507849 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:12 crc kubenswrapper[4871]: I1007 22:09:12.546931 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:12Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.247562 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xwhnv" event={"ID":"d535a1fa-1557-4e32-8583-d8415c47c1bf","Type":"ContainerStarted","Data":"7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.254091 4871 generic.go:334] "Generic (PLEG): container finished" podID="ad179ea2-56e8-4e5e-a72c-599c797918d1" containerID="018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad" exitCode=0 Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.254211 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" event={"ID":"ad179ea2-56e8-4e5e-a72c-599c797918d1","Type":"ContainerDied","Data":"018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.268415 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.297919 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.336600 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.361126 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.388009 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.411146 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.432598 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.452932 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.474682 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.485177 4871 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.488104 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.488158 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.488173 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.488353 4871 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.493273 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.498823 4871 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.499206 4871 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.500417 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.500448 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.500459 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.500589 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.500619 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:13Z","lastTransitionTime":"2025-10-07T22:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.511390 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: E1007 22:09:13.516029 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.521707 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.521860 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.521924 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.521987 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.522066 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:13Z","lastTransitionTime":"2025-10-07T22:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.530468 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: E1007 22:09:13.536383 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.540972 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.541052 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.541071 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.541098 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.541116 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:13Z","lastTransitionTime":"2025-10-07T22:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.550499 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: E1007 22:09:13.559035 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.564389 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.564447 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.564462 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.564482 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.564497 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:13Z","lastTransitionTime":"2025-10-07T22:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.565415 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.582466 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: E1007 22:09:13.584246 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.589227 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.589280 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.589292 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.589342 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.589361 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:13Z","lastTransitionTime":"2025-10-07T22:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.600864 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: E1007 22:09:13.608424 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: E1007 22:09:13.608889 4871 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.612274 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.612321 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.612334 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.612351 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.612364 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:13Z","lastTransitionTime":"2025-10-07T22:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.624667 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.649015 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.673196 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.690169 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.708927 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.715070 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.715115 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.715130 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.715153 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.715172 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:13Z","lastTransitionTime":"2025-10-07T22:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.730229 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.750287 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.767874 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.789315 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.812672 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.819163 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.819240 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.819255 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.819281 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.819295 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:13Z","lastTransitionTime":"2025-10-07T22:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.835935 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.858984 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:13Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.923211 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.923290 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.923312 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.923342 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.923361 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:13Z","lastTransitionTime":"2025-10-07T22:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.982374 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.982434 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:13 crc kubenswrapper[4871]: I1007 22:09:13.982403 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:13 crc kubenswrapper[4871]: E1007 22:09:13.982596 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:13 crc kubenswrapper[4871]: E1007 22:09:13.982755 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:13 crc kubenswrapper[4871]: E1007 22:09:13.982892 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.026573 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.026625 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.026636 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.026653 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.026667 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.130178 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.130238 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.130255 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.130281 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.130296 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.233859 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.233923 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.233942 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.233970 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.233989 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.265143 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.268627 4871 generic.go:334] "Generic (PLEG): container finished" podID="ad179ea2-56e8-4e5e-a72c-599c797918d1" containerID="95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1" exitCode=0 Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.268735 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" event={"ID":"ad179ea2-56e8-4e5e-a72c-599c797918d1","Type":"ContainerDied","Data":"95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.292235 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.314529 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.331712 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.338181 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.338240 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.338261 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.338288 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.338307 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.346318 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.367813 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.380514 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.403562 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.426036 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.441893 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.441945 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.441957 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.441976 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.441990 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.444210 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.462503 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.482336 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.497039 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.513757 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.527922 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:14Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.545842 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.545891 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.545905 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.545925 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.545940 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.649518 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.649600 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.649622 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.649652 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.649674 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.753221 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.753311 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.753331 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.753361 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.753378 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.856627 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.856705 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.856751 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.856784 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.856823 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.960234 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.960345 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.960369 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.960402 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:14 crc kubenswrapper[4871]: I1007 22:09:14.960424 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:14Z","lastTransitionTime":"2025-10-07T22:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.063954 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.064023 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.064042 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.064071 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.064090 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:15Z","lastTransitionTime":"2025-10-07T22:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.167616 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.167678 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.167692 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.167714 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.167728 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:15Z","lastTransitionTime":"2025-10-07T22:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.272373 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.272423 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.272435 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.272458 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.272476 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:15Z","lastTransitionTime":"2025-10-07T22:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.277332 4871 generic.go:334] "Generic (PLEG): container finished" podID="ad179ea2-56e8-4e5e-a72c-599c797918d1" containerID="029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13" exitCode=0 Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.277370 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" event={"ID":"ad179ea2-56e8-4e5e-a72c-599c797918d1","Type":"ContainerDied","Data":"029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.296460 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.313985 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.330390 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.348881 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.356753 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.357465 4871 scope.go:117] "RemoveContainer" containerID="86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23" Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.357650 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.366559 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.375584 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.375630 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.375639 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.375657 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.375671 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:15Z","lastTransitionTime":"2025-10-07T22:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.382295 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.408111 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.443357 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.460994 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.481610 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.483264 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.483316 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.483334 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.483357 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.483372 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:15Z","lastTransitionTime":"2025-10-07T22:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.506282 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.525641 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.545388 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.577557 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.587049 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.587113 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.587135 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.587165 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.587184 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:15Z","lastTransitionTime":"2025-10-07T22:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.592832 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.592972 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.593041 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.593104 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.593162 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.593352 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.593395 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.593417 4871 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.593495 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:23.593466171 +0000 UTC m=+37.396164284 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.594189 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:09:23.594128639 +0000 UTC m=+37.396826752 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.594390 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.594468 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.594487 4871 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.594581 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:23.5945624 +0000 UTC m=+37.397260523 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.594746 4871 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.594874 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:23.594850848 +0000 UTC m=+37.397548961 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.595070 4871 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.595174 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:23.595157656 +0000 UTC m=+37.397855769 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.689561 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.689600 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.689626 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.689644 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.689657 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:15Z","lastTransitionTime":"2025-10-07T22:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.793500 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.793539 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.793553 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.793572 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.793587 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:15Z","lastTransitionTime":"2025-10-07T22:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.897560 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.897604 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.897617 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.897635 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.897646 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:15Z","lastTransitionTime":"2025-10-07T22:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.982226 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.982444 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.982570 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:15 crc kubenswrapper[4871]: I1007 22:09:15.982713 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.982761 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:15 crc kubenswrapper[4871]: E1007 22:09:15.983027 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.001113 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.001200 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.001228 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.001266 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.001293 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.104656 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.104718 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.104731 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.104754 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.104769 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.207030 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.207073 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.207084 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.207101 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.207114 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.287262 4871 generic.go:334] "Generic (PLEG): container finished" podID="ad179ea2-56e8-4e5e-a72c-599c797918d1" containerID="b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5" exitCode=0 Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.287397 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" event={"ID":"ad179ea2-56e8-4e5e-a72c-599c797918d1","Type":"ContainerDied","Data":"b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.297988 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.298401 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.310226 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.310344 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.310377 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.310413 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.310439 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.314233 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.334042 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.342123 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.351989 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.368333 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.384104 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.408090 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.413520 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.413559 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.413573 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.413593 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.413605 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.441924 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.465968 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.489838 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.508553 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.517369 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.517473 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.517494 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.517530 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.517551 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.528588 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.550439 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.570383 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.590354 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.607750 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.620736 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.620858 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.620888 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.620922 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.620946 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.621746 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.635953 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.655241 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.667047 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.680037 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.698176 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.715073 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.724995 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.725066 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.725086 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.725115 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.725141 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.731650 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.745993 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.764453 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.778336 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.796601 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.819320 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.828920 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.829082 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.829167 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.829262 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.829355 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.932960 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.933029 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.933047 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.933075 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.933092 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:16Z","lastTransitionTime":"2025-10-07T22:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:16 crc kubenswrapper[4871]: I1007 22:09:16.999356 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:16Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.011437 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.030840 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.036369 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.036439 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.036458 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.036485 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.036504 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.062430 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.081925 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.104899 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.106906 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.126170 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.139245 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.139292 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.139304 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.139349 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.139362 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.147236 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.164486 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.179667 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.193054 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.206370 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.220690 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.236493 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.242210 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.242262 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.242285 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.242313 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.242330 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.308203 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" event={"ID":"ad179ea2-56e8-4e5e-a72c-599c797918d1","Type":"ContainerStarted","Data":"6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.309522 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.327604 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.339263 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.343631 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.345849 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.346024 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.346215 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.346470 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.347327 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.360427 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.381321 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.434382 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.450335 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.450375 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.450388 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.450410 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.450424 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.460605 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.482032 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.496182 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.512578 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.524419 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.538552 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.553466 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.553504 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.553516 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.553535 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.553546 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.554066 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.580012 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.598831 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.616663 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.630095 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.644816 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.667660 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.668008 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.668094 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.668181 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.668253 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.672461 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.685818 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.699753 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.713487 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.732065 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.752529 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.767524 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.771340 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.771563 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.771688 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.771785 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.771886 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.780063 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.793243 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.806615 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.819218 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.875529 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.875946 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.876225 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.876468 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.876721 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.979729 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.980046 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.980121 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.980208 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.980280 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:17Z","lastTransitionTime":"2025-10-07T22:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.981425 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.981531 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:17 crc kubenswrapper[4871]: E1007 22:09:17.981628 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:17 crc kubenswrapper[4871]: I1007 22:09:17.981433 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:17 crc kubenswrapper[4871]: E1007 22:09:17.981749 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:17 crc kubenswrapper[4871]: E1007 22:09:17.981782 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.084147 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.084533 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.084649 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.084774 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.084871 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:18Z","lastTransitionTime":"2025-10-07T22:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.187859 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.187910 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.187922 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.187941 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.187953 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:18Z","lastTransitionTime":"2025-10-07T22:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.290685 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.290734 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.290745 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.290762 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.290778 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:18Z","lastTransitionTime":"2025-10-07T22:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.394066 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.394141 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.394159 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.394189 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.394212 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:18Z","lastTransitionTime":"2025-10-07T22:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.497136 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.497172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.497183 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.497199 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.497210 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:18Z","lastTransitionTime":"2025-10-07T22:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.599901 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.599941 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.599953 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.599976 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.599991 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:18Z","lastTransitionTime":"2025-10-07T22:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.703910 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.703972 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.703986 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.704008 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.704023 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:18Z","lastTransitionTime":"2025-10-07T22:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.807975 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.808050 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.808069 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.808097 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.808118 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:18Z","lastTransitionTime":"2025-10-07T22:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.911958 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.912032 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.912051 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.912079 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:18 crc kubenswrapper[4871]: I1007 22:09:18.912100 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:18Z","lastTransitionTime":"2025-10-07T22:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.014767 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.014845 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.014858 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.014874 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.014887 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.118397 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.118470 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.118488 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.118516 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.118537 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.222212 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.222277 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.222295 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.222322 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.222341 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.320639 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/0.log" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.324584 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.324635 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.324653 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.324679 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.324698 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.325142 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036" exitCode=1 Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.325204 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.326290 4871 scope.go:117] "RemoveContainer" containerID="e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.348543 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.372744 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.390005 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.404740 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.424438 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.428117 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.428162 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.428176 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.428198 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.428212 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.448015 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.463781 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.484110 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.509823 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.531263 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.531329 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.531348 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.531379 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.531399 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.546326 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:18Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:18.907593 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 22:09:18.907717 6218 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 22:09:18.907757 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:18.907864 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 22:09:18.907876 6218 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:18.907902 6218 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 22:09:18.907910 6218 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:18.907913 6218 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:18.907966 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 22:09:18.907983 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 22:09:18.908033 6218 factory.go:656] Stopping watch factory\\\\nI1007 22:09:18.908040 6218 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:18.908062 6218 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:18.908065 6218 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:18.908115 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.571610 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.598493 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.618644 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.635423 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.635494 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.635512 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.635540 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.635557 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.636157 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:19Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.739987 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.740072 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.740091 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.740120 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.740141 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.844388 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.844516 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.844545 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.844579 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.844608 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.947754 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.947864 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.947887 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.947916 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.947938 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:19Z","lastTransitionTime":"2025-10-07T22:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.981600 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.981644 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:19 crc kubenswrapper[4871]: I1007 22:09:19.981738 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:19 crc kubenswrapper[4871]: E1007 22:09:19.981782 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:19 crc kubenswrapper[4871]: E1007 22:09:19.981988 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:19 crc kubenswrapper[4871]: E1007 22:09:19.982108 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.051362 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.051419 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.051433 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.051455 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.051471 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.154522 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.154583 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.154597 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.154619 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.154636 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.258363 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.258429 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.258448 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.258478 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.258498 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.333593 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/0.log" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.337527 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.338165 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.356037 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.362696 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.362753 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.362773 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.362835 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.362855 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.376277 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.395354 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.426208 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.441230 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.458688 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.467292 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.467357 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.467377 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.467405 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.467425 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.473731 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.497253 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:18Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:18.907593 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 22:09:18.907717 6218 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 22:09:18.907757 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:18.907864 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 22:09:18.907876 6218 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:18.907902 6218 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 22:09:18.907910 6218 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:18.907913 6218 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:18.907966 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 22:09:18.907983 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 22:09:18.908033 6218 factory.go:656] Stopping watch factory\\\\nI1007 22:09:18.908040 6218 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:18.908062 6218 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:18.908065 6218 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:18.908115 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.511435 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.522538 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.538399 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.549919 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.565146 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.569860 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.569905 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.569923 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.569949 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.569968 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.580950 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:20Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.674282 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.674364 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.674382 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.674409 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.674429 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.777715 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.777779 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.777827 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.777853 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.777874 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.881074 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.881355 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.881507 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.881696 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.881953 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.986262 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.986340 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.986359 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.986392 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:20 crc kubenswrapper[4871]: I1007 22:09:20.986416 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:20Z","lastTransitionTime":"2025-10-07T22:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.090169 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.090675 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.090880 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.091065 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.091206 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:21Z","lastTransitionTime":"2025-10-07T22:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.194920 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.194991 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.195012 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.195041 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.195065 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:21Z","lastTransitionTime":"2025-10-07T22:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.298983 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.299418 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.299441 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.299468 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.299485 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:21Z","lastTransitionTime":"2025-10-07T22:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.346073 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/1.log" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.347016 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/0.log" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.352385 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209" exitCode=1 Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.352442 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.352499 4871 scope.go:117] "RemoveContainer" containerID="e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.353635 4871 scope.go:117] "RemoveContainer" containerID="206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209" Oct 07 22:09:21 crc kubenswrapper[4871]: E1007 22:09:21.353973 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.378197 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.405999 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.406894 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.406970 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.406997 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.407028 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.407051 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:21Z","lastTransitionTime":"2025-10-07T22:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.429325 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.447001 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.474375 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8"] Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.475233 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.476340 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.479591 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.480167 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.499588 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.510474 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.510533 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.510550 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.510589 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.510605 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:21Z","lastTransitionTime":"2025-10-07T22:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.535593 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:18Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:18.907593 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 22:09:18.907717 6218 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 22:09:18.907757 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:18.907864 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 22:09:18.907876 6218 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:18.907902 6218 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 22:09:18.907910 6218 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:18.907913 6218 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:18.907966 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 22:09:18.907983 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 22:09:18.908033 6218 factory.go:656] Stopping watch factory\\\\nI1007 22:09:18.908040 6218 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:18.908062 6218 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:18.908065 6218 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:18.908115 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.556136 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.579982 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.601166 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.601969 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d5008be0-9415-428f-917e-7ec0905be00d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.602093 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkz4b\" (UniqueName: \"kubernetes.io/projected/d5008be0-9415-428f-917e-7ec0905be00d-kube-api-access-xkz4b\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.602317 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5008be0-9415-428f-917e-7ec0905be00d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.602357 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d5008be0-9415-428f-917e-7ec0905be00d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.613610 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.613678 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.613701 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.613731 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.613755 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:21Z","lastTransitionTime":"2025-10-07T22:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.622237 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.643626 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.663686 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.685552 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.704175 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5008be0-9415-428f-917e-7ec0905be00d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.704291 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d5008be0-9415-428f-917e-7ec0905be00d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.704476 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d5008be0-9415-428f-917e-7ec0905be00d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.704569 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkz4b\" (UniqueName: \"kubernetes.io/projected/d5008be0-9415-428f-917e-7ec0905be00d-kube-api-access-xkz4b\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.705870 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5008be0-9415-428f-917e-7ec0905be00d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.706202 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d5008be0-9415-428f-917e-7ec0905be00d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.708635 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.715934 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d5008be0-9415-428f-917e-7ec0905be00d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.717781 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.717851 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.717868 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.717897 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.717919 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:21Z","lastTransitionTime":"2025-10-07T22:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.731787 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.738594 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkz4b\" (UniqueName: \"kubernetes.io/projected/d5008be0-9415-428f-917e-7ec0905be00d-kube-api-access-xkz4b\") pod \"ovnkube-control-plane-749d76644c-hlhw8\" (UID: \"d5008be0-9415-428f-917e-7ec0905be00d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.751904 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.774075 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.793886 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.802396 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.817293 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.820453 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.820704 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.821099 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.821335 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.821506 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:21Z","lastTransitionTime":"2025-10-07T22:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: W1007 22:09:21.827253 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5008be0_9415_428f_917e_7ec0905be00d.slice/crio-a9466153e7892e124f6676b052f48575ecce356921fd97bc23407c3a827ba006 WatchSource:0}: Error finding container a9466153e7892e124f6676b052f48575ecce356921fd97bc23407c3a827ba006: Status 404 returned error can't find the container with id a9466153e7892e124f6676b052f48575ecce356921fd97bc23407c3a827ba006 Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.834664 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.849692 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.873664 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.896749 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3f0bb263eeafd1f6b214cea104a0a330eb6f8f2dbf086646b64600f2f7e0036\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:18Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:18.907593 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 22:09:18.907717 6218 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 22:09:18.907757 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:18.907864 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 22:09:18.907876 6218 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:18.907902 6218 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 22:09:18.907910 6218 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:18.907913 6218 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:18.907966 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 22:09:18.907983 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 22:09:18.908033 6218 factory.go:656] Stopping watch factory\\\\nI1007 22:09:18.908040 6218 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:18.908062 6218 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:18.908065 6218 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:18.908115 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.910634 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.925160 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.925199 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.925211 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.925230 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.925243 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:21Z","lastTransitionTime":"2025-10-07T22:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.926227 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.952420 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.968611 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.981438 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:21 crc kubenswrapper[4871]: E1007 22:09:21.981609 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.981628 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.981732 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:21 crc kubenswrapper[4871]: E1007 22:09:21.981749 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:21 crc kubenswrapper[4871]: E1007 22:09:21.981956 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:21 crc kubenswrapper[4871]: I1007 22:09:21.982827 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:21Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.028475 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.028510 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.028523 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.028543 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.028556 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.131495 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.131562 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.131578 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.131601 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.131614 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.234820 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.234853 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.234862 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.234878 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.234887 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.337422 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.337479 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.337496 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.337520 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.337537 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.357931 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/1.log" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.362563 4871 scope.go:117] "RemoveContainer" containerID="206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.362734 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" event={"ID":"d5008be0-9415-428f-917e-7ec0905be00d","Type":"ContainerStarted","Data":"8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.362784 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" event={"ID":"d5008be0-9415-428f-917e-7ec0905be00d","Type":"ContainerStarted","Data":"3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.362813 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" event={"ID":"d5008be0-9415-428f-917e-7ec0905be00d","Type":"ContainerStarted","Data":"a9466153e7892e124f6676b052f48575ecce356921fd97bc23407c3a827ba006"} Oct 07 22:09:22 crc kubenswrapper[4871]: E1007 22:09:22.362958 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.376668 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.389182 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.400285 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.408827 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.420884 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.435955 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.440203 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.440442 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.440554 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.440631 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.440699 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.458014 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.480562 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.495424 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.513383 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.525850 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.537837 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.543163 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.543216 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.543229 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.543249 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.543265 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.550962 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.563615 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.574733 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.592494 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.604365 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.615023 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.627669 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.640043 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.646063 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.646122 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.646141 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.646168 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.646186 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.654831 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.666693 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.678684 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.689292 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.706270 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.729351 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.747302 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.750172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.750252 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.750271 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.750328 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.750349 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.761893 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.775449 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.790626 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:22Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.854698 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.854747 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.854758 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.854776 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.854811 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.958284 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.958369 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.958388 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.958414 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.958435 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:22Z","lastTransitionTime":"2025-10-07T22:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.995443 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-gbxdg"] Oct 07 22:09:22 crc kubenswrapper[4871]: I1007 22:09:22.996367 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:22 crc kubenswrapper[4871]: E1007 22:09:22.996480 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.021309 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.042516 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.061778 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.061873 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.061896 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.061924 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.061943 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.064209 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.082511 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.102593 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.118923 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.119222 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xsct\" (UniqueName: \"kubernetes.io/projected/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-kube-api-access-9xsct\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.122654 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.145301 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.163249 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.165464 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.165515 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.165536 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.165568 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.165588 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.180885 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.205245 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.220453 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xsct\" (UniqueName: \"kubernetes.io/projected/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-kube-api-access-9xsct\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.220590 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.220895 4871 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.221004 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs podName:684fd639-9051-4c1a-8bf0-9d1cb7b5cf59 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:23.720975818 +0000 UTC m=+37.523673931 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs") pod "network-metrics-daemon-gbxdg" (UID: "684fd639-9051-4c1a-8bf0-9d1cb7b5cf59") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.237474 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.255045 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xsct\" (UniqueName: \"kubernetes.io/projected/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-kube-api-access-9xsct\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.263698 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.269419 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.269483 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.269509 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.269542 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.269569 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.286664 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.309031 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.331515 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.351511 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.373119 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.373187 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.373210 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.373241 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.373270 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.478351 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.478422 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.478442 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.478471 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.478489 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.582630 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.582711 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.582734 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.582767 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.582861 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.625290 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.625440 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.625529 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.625554 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:09:39.625515349 +0000 UTC m=+53.428213472 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.625597 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.625727 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.625777 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.625809 4871 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.625731 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.625831 4871 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.625892 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:39.625868938 +0000 UTC m=+53.428567011 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.625727 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.625937 4871 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.625959 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:39.62593551 +0000 UTC m=+53.428633793 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.626008 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.626017 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:39.625994862 +0000 UTC m=+53.428693135 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.626030 4871 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.626105 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:39.626086234 +0000 UTC m=+53.428784337 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.686371 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.686421 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.686439 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.686463 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.686483 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.726577 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.726847 4871 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.726952 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs podName:684fd639-9051-4c1a-8bf0-9d1cb7b5cf59 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:24.726927214 +0000 UTC m=+38.529625287 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs") pod "network-metrics-daemon-gbxdg" (UID: "684fd639-9051-4c1a-8bf0-9d1cb7b5cf59") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.791033 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.791132 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.791153 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.791185 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.791214 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.840173 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.840250 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.840276 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.840307 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.840327 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.861573 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.866988 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.867055 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.867076 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.867102 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.867121 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.891188 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.897127 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.897172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.897184 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.897203 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.897218 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.915686 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.921561 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.921607 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.921621 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.921640 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.921656 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.942890 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.948490 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.948581 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.948601 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.948630 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.948649 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.971965 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:23Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.972219 4871 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.974760 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.974821 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.974832 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.974855 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.974866 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:23Z","lastTransitionTime":"2025-10-07T22:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.981935 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.981981 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.982056 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.982156 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:23 crc kubenswrapper[4871]: I1007 22:09:23.982234 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:23 crc kubenswrapper[4871]: E1007 22:09:23.982455 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.078546 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.078646 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.078672 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.078708 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.078734 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:24Z","lastTransitionTime":"2025-10-07T22:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.183119 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.183182 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.183205 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.183236 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.183257 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:24Z","lastTransitionTime":"2025-10-07T22:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.286750 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.286857 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.286879 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.286907 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.286928 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:24Z","lastTransitionTime":"2025-10-07T22:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.390977 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.391042 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.391060 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.391086 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.391105 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:24Z","lastTransitionTime":"2025-10-07T22:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.495530 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.495922 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.495967 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.496010 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.496056 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:24Z","lastTransitionTime":"2025-10-07T22:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.606055 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.606157 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.606169 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.606191 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.606213 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:24Z","lastTransitionTime":"2025-10-07T22:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.709847 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.709896 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.709906 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.709926 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.709938 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:24Z","lastTransitionTime":"2025-10-07T22:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.738750 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:24 crc kubenswrapper[4871]: E1007 22:09:24.738972 4871 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:24 crc kubenswrapper[4871]: E1007 22:09:24.739104 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs podName:684fd639-9051-4c1a-8bf0-9d1cb7b5cf59 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:26.739078542 +0000 UTC m=+40.541776655 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs") pod "network-metrics-daemon-gbxdg" (UID: "684fd639-9051-4c1a-8bf0-9d1cb7b5cf59") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.813187 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.813278 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.813298 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.813329 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.813350 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:24Z","lastTransitionTime":"2025-10-07T22:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.917439 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.917540 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.917570 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.917609 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.917636 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:24Z","lastTransitionTime":"2025-10-07T22:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:24 crc kubenswrapper[4871]: I1007 22:09:24.982547 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:24 crc kubenswrapper[4871]: E1007 22:09:24.982863 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.020917 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.020993 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.021016 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.021048 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.021068 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.125172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.125259 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.125276 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.125306 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.125325 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.229208 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.229280 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.229298 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.229327 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.229346 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.333236 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.333309 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.333330 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.333359 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.333382 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.437210 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.437270 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.437284 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.437308 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.437322 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.541320 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.541994 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.542038 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.542066 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.542082 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.645507 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.645574 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.645598 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.645624 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.645647 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.749619 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.749681 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.749702 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.749728 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.749748 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.853241 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.853319 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.853342 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.853375 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.853398 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.956597 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.957201 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.957444 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.957673 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.957914 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:25Z","lastTransitionTime":"2025-10-07T22:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.982004 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.982052 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:25 crc kubenswrapper[4871]: I1007 22:09:25.982624 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:25 crc kubenswrapper[4871]: E1007 22:09:25.982928 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:25 crc kubenswrapper[4871]: E1007 22:09:25.982787 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:25 crc kubenswrapper[4871]: E1007 22:09:25.983459 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.062473 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.062531 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.062544 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.062564 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.062578 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.166248 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.166344 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.166358 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.166380 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.166393 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.269437 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.269488 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.269499 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.269526 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.269539 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.373296 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.373365 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.373383 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.373415 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.373435 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.477056 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.477112 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.477126 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.477142 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.477152 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.581095 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.581190 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.581217 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.581251 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.581273 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.685280 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.685356 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.685380 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.685412 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.685435 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.764360 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:26 crc kubenswrapper[4871]: E1007 22:09:26.764642 4871 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:26 crc kubenswrapper[4871]: E1007 22:09:26.764734 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs podName:684fd639-9051-4c1a-8bf0-9d1cb7b5cf59 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:30.764706591 +0000 UTC m=+44.567404694 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs") pod "network-metrics-daemon-gbxdg" (UID: "684fd639-9051-4c1a-8bf0-9d1cb7b5cf59") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.789739 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.789907 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.789933 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.789969 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.789990 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.893001 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.893429 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.893527 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.893640 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.893732 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.981838 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:26 crc kubenswrapper[4871]: E1007 22:09:26.982122 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.982734 4871 scope.go:117] "RemoveContainer" containerID="86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.996633 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.996886 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.997029 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.997249 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:26 crc kubenswrapper[4871]: I1007 22:09:26.997409 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:26Z","lastTransitionTime":"2025-10-07T22:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.012690 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.030663 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.052066 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.075058 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.092741 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.100706 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.101034 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.101058 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.101080 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.101095 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:27Z","lastTransitionTime":"2025-10-07T22:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.109717 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.128898 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.149498 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.179537 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.197195 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.204580 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.204637 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.204654 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.204677 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.204690 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:27Z","lastTransitionTime":"2025-10-07T22:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.211459 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.232825 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.247628 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.264853 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.290260 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.307586 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.307662 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.307677 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.307723 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.307737 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:27Z","lastTransitionTime":"2025-10-07T22:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.313154 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.385226 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.388209 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.388603 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.407333 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.410537 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.410590 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.410611 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.410637 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.410657 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:27Z","lastTransitionTime":"2025-10-07T22:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.420534 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.443013 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.481848 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.510146 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.514395 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.514445 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.514460 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.514478 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.514491 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:27Z","lastTransitionTime":"2025-10-07T22:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.534899 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.553128 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.570982 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.591585 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.613495 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.617219 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.617266 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.617283 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.617308 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.617328 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:27Z","lastTransitionTime":"2025-10-07T22:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.632713 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.649891 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.667695 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.688276 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.708660 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.720994 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.721084 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.721111 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.721565 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.721879 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:27Z","lastTransitionTime":"2025-10-07T22:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.727586 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.825469 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.825579 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.825634 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.825662 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.825681 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:27Z","lastTransitionTime":"2025-10-07T22:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.929611 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.929680 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.929703 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.929731 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.929749 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:27Z","lastTransitionTime":"2025-10-07T22:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.981282 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:27 crc kubenswrapper[4871]: E1007 22:09:27.981476 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.982082 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:27 crc kubenswrapper[4871]: E1007 22:09:27.982195 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:27 crc kubenswrapper[4871]: I1007 22:09:27.982276 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:27 crc kubenswrapper[4871]: E1007 22:09:27.982355 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.034993 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.035361 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.035496 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.035693 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.035862 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.139718 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.139832 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.139861 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.139892 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.139918 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.243871 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.243969 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.243998 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.244033 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.244058 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.347733 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.347841 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.347867 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.347896 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.347916 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.452040 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.452476 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.452647 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.452834 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.452974 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.556967 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.557500 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.557652 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.557850 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.558017 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.661292 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.661372 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.661394 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.661425 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.661447 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.764612 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.764693 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.764717 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.764749 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.764775 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.867827 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.867919 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.867944 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.867974 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.867994 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.971877 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.971929 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.971947 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.971967 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.971980 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:28Z","lastTransitionTime":"2025-10-07T22:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:28 crc kubenswrapper[4871]: I1007 22:09:28.982013 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:28 crc kubenswrapper[4871]: E1007 22:09:28.982258 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.075833 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.075900 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.075915 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.075940 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.075957 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:29Z","lastTransitionTime":"2025-10-07T22:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.180153 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.180234 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.180248 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.180266 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.180285 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:29Z","lastTransitionTime":"2025-10-07T22:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.283351 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.283424 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.283441 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.283465 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.283486 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:29Z","lastTransitionTime":"2025-10-07T22:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.387381 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.387431 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.387442 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.387461 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.387473 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:29Z","lastTransitionTime":"2025-10-07T22:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.491232 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.491293 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.491316 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.491346 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.491368 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:29Z","lastTransitionTime":"2025-10-07T22:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.595347 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.595697 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.595923 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.596073 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.596233 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:29Z","lastTransitionTime":"2025-10-07T22:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.699690 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.700678 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.700870 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.701043 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.701189 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:29Z","lastTransitionTime":"2025-10-07T22:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.804664 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.804750 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.804775 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.804847 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.804873 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:29Z","lastTransitionTime":"2025-10-07T22:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.909272 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.909344 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.909363 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.909389 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.909408 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:29Z","lastTransitionTime":"2025-10-07T22:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.981280 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.981315 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:29 crc kubenswrapper[4871]: I1007 22:09:29.981429 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:29 crc kubenswrapper[4871]: E1007 22:09:29.981494 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:29 crc kubenswrapper[4871]: E1007 22:09:29.981615 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:29 crc kubenswrapper[4871]: E1007 22:09:29.981724 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.012705 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.012772 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.012821 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.012857 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.012878 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.116302 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.116382 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.116399 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.116434 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.116454 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.220180 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.220249 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.220271 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.220298 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.220318 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.324224 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.324298 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.324316 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.324345 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.324367 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.428883 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.428966 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.428985 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.429018 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.429040 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.532393 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.532471 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.532489 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.532516 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.532534 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.635847 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.635916 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.635935 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.635963 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.635982 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.739529 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.739592 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.739613 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.739641 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.739665 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.811947 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:30 crc kubenswrapper[4871]: E1007 22:09:30.812186 4871 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:30 crc kubenswrapper[4871]: E1007 22:09:30.812297 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs podName:684fd639-9051-4c1a-8bf0-9d1cb7b5cf59 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:38.812265705 +0000 UTC m=+52.614963808 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs") pod "network-metrics-daemon-gbxdg" (UID: "684fd639-9051-4c1a-8bf0-9d1cb7b5cf59") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.842865 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.842940 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.842963 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.842996 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.843021 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.946281 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.946366 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.946386 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.946414 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.946434 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:30Z","lastTransitionTime":"2025-10-07T22:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:30 crc kubenswrapper[4871]: I1007 22:09:30.981649 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:30 crc kubenswrapper[4871]: E1007 22:09:30.981902 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.049766 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.049862 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.049884 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.049905 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.049921 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.154011 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.154096 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.154116 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.154144 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.154166 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.257858 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.257923 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.257939 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.257964 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.257980 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.361767 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.361867 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.361886 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.361913 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.361932 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.466217 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.466289 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.466310 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.466340 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.466361 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.569932 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.569984 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.569996 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.570014 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.570027 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.673349 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.673421 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.673448 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.673481 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.673504 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.777245 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.777314 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.777332 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.777361 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.777384 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.881076 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.881142 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.881162 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.881189 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.881207 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.981821 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.981898 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:31 crc kubenswrapper[4871]: E1007 22:09:31.982095 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:31 crc kubenswrapper[4871]: E1007 22:09:31.982224 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.981869 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:31 crc kubenswrapper[4871]: E1007 22:09:31.982770 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.984964 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.985011 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.985029 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.985052 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:31 crc kubenswrapper[4871]: I1007 22:09:31.985070 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:31Z","lastTransitionTime":"2025-10-07T22:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.088317 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.088670 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.088700 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.088728 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.088747 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:32Z","lastTransitionTime":"2025-10-07T22:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.192528 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.192589 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.192608 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.192633 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.192650 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:32Z","lastTransitionTime":"2025-10-07T22:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.295835 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.295890 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.295906 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.295930 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.295947 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:32Z","lastTransitionTime":"2025-10-07T22:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.399108 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.399160 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.399176 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.399200 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.399217 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:32Z","lastTransitionTime":"2025-10-07T22:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.503037 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.503119 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.503141 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.503165 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.503183 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:32Z","lastTransitionTime":"2025-10-07T22:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.606057 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.606106 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.606116 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.606135 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.606150 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:32Z","lastTransitionTime":"2025-10-07T22:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.709481 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.709537 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.709554 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.709585 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.709601 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:32Z","lastTransitionTime":"2025-10-07T22:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.813351 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.813419 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.813436 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.813463 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.813481 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:32Z","lastTransitionTime":"2025-10-07T22:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.917998 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.918063 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.918079 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.918103 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.918121 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:32Z","lastTransitionTime":"2025-10-07T22:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:32 crc kubenswrapper[4871]: I1007 22:09:32.982369 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:32 crc kubenswrapper[4871]: E1007 22:09:32.982604 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.022016 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.022350 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.022377 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.022593 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.022618 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.126328 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.126410 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.126435 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.126471 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.126498 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.230913 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.230982 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.230999 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.231026 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.231043 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.334856 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.334920 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.334938 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.334963 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.334982 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.439066 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.439163 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.439192 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.439251 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.439278 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.543169 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.543219 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.543228 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.543246 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.543256 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.649003 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.649059 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.649075 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.649097 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.649110 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.752499 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.752552 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.752565 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.752587 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.752600 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.856343 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.856418 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.856439 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.856466 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.856485 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.960636 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.960705 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.960722 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.960756 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.960777 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:33Z","lastTransitionTime":"2025-10-07T22:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.982171 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.982210 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:33 crc kubenswrapper[4871]: I1007 22:09:33.982169 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:33 crc kubenswrapper[4871]: E1007 22:09:33.982373 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:33 crc kubenswrapper[4871]: E1007 22:09:33.982565 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:33 crc kubenswrapper[4871]: E1007 22:09:33.982747 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.030829 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.030901 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.030920 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.030948 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.030970 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: E1007 22:09:34.052485 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:34Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.059394 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.059460 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.059491 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.059527 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.059547 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: E1007 22:09:34.081117 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:34Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.089204 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.089273 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.089291 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.089319 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.089336 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: E1007 22:09:34.109920 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:34Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.116311 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.116386 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.116409 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.116441 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.116465 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: E1007 22:09:34.137028 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:34Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.142845 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.142900 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.142918 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.142947 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.142966 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: E1007 22:09:34.164223 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:34Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:34 crc kubenswrapper[4871]: E1007 22:09:34.164459 4871 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.167228 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.167284 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.167304 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.167333 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.167354 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.272701 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.272827 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.272912 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.272956 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.273108 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.377110 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.377196 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.377220 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.377249 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.377266 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.480835 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.480916 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.480935 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.480963 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.480985 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.584228 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.584299 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.584316 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.584344 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.584365 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.688026 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.688098 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.688116 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.688147 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.688168 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.792708 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.792778 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.792819 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.792849 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.792868 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.896178 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.896232 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.896245 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.896265 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.896278 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.982381 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:34 crc kubenswrapper[4871]: E1007 22:09:34.982640 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.999164 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.999210 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.999222 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.999239 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:34 crc kubenswrapper[4871]: I1007 22:09:34.999251 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:34Z","lastTransitionTime":"2025-10-07T22:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.103269 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.103349 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.103374 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.103406 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.103428 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:35Z","lastTransitionTime":"2025-10-07T22:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.208187 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.208257 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.208280 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.208307 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.208330 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:35Z","lastTransitionTime":"2025-10-07T22:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.312264 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.312409 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.312430 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.312460 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.312479 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:35Z","lastTransitionTime":"2025-10-07T22:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.417076 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.417142 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.417160 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.417217 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.417237 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:35Z","lastTransitionTime":"2025-10-07T22:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.520696 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.520778 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.520841 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.520875 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.520898 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:35Z","lastTransitionTime":"2025-10-07T22:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.624208 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.624261 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.624276 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.624295 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.624311 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:35Z","lastTransitionTime":"2025-10-07T22:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.727516 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.727562 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.727573 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.727591 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.727605 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:35Z","lastTransitionTime":"2025-10-07T22:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.831068 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.831110 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.831120 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.831136 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.831147 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:35Z","lastTransitionTime":"2025-10-07T22:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.934172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.934227 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.934240 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.934261 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.934274 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:35Z","lastTransitionTime":"2025-10-07T22:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.981294 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.981417 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:35 crc kubenswrapper[4871]: I1007 22:09:35.981294 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:35 crc kubenswrapper[4871]: E1007 22:09:35.981577 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:35 crc kubenswrapper[4871]: E1007 22:09:35.981717 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:35 crc kubenswrapper[4871]: E1007 22:09:35.982001 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.037208 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.037248 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.037260 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.037281 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.037291 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.141177 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.141254 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.141280 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.141313 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.141338 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.245174 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.245240 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.245262 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.245297 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.245317 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.348744 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.348938 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.348964 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.349030 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.349054 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.452564 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.452638 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.452655 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.452689 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.452708 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.555854 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.555967 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.555987 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.556015 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.556035 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.659925 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.659991 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.660010 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.660038 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.660059 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.763980 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.764032 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.764051 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.764072 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.764091 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.867740 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.867889 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.867908 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.867934 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.867983 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.972693 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.972753 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.972772 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.972831 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.972851 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:36Z","lastTransitionTime":"2025-10-07T22:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.983384 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:36 crc kubenswrapper[4871]: I1007 22:09:36.984523 4871 scope.go:117] "RemoveContainer" containerID="206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209" Oct 07 22:09:36 crc kubenswrapper[4871]: E1007 22:09:36.985171 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.007746 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.026750 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.054908 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.077835 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.077882 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.077898 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.077925 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.077946 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:37Z","lastTransitionTime":"2025-10-07T22:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.091778 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.112609 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.135858 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.162612 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.182017 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.182340 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.182346 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.182652 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.182708 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.182727 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:37Z","lastTransitionTime":"2025-10-07T22:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.200053 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.222635 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.245510 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.266408 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.279507 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.285677 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.285747 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.285769 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.285823 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.285845 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:37Z","lastTransitionTime":"2025-10-07T22:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.301751 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.323583 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.339019 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.389087 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.389129 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.389140 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.389157 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.389170 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:37Z","lastTransitionTime":"2025-10-07T22:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.429543 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/1.log" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.434398 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.435759 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.454667 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.475369 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.495222 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.495299 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.495313 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.495347 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.495362 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:37Z","lastTransitionTime":"2025-10-07T22:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.502035 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.518520 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.537405 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.559735 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.583101 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.597983 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.598029 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.598040 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.598059 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.598077 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:37Z","lastTransitionTime":"2025-10-07T22:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.613261 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.641735 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.671129 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.689659 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.700773 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.700850 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.700866 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.700891 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.700906 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:37Z","lastTransitionTime":"2025-10-07T22:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.705206 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.720317 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.735469 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.748412 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.761308 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:37Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.804009 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.804068 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.804082 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.804105 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.804120 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:37Z","lastTransitionTime":"2025-10-07T22:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.906963 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.907029 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.907050 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.907076 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.907095 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:37Z","lastTransitionTime":"2025-10-07T22:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.981769 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.981884 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:37 crc kubenswrapper[4871]: I1007 22:09:37.981960 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:37 crc kubenswrapper[4871]: E1007 22:09:37.982022 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:37 crc kubenswrapper[4871]: E1007 22:09:37.982134 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:37 crc kubenswrapper[4871]: E1007 22:09:37.982243 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.009881 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.009945 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.009963 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.009998 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.010018 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.113083 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.113143 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.113161 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.113188 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.113206 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.215586 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.215628 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.215637 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.215651 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.215660 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.319423 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.319524 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.319544 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.319601 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.319622 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.423067 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.423169 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.423188 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.423215 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.423235 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.441556 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/2.log" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.442712 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/1.log" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.447761 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11" exitCode=1 Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.447887 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.448006 4871 scope.go:117] "RemoveContainer" containerID="206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.449262 4871 scope.go:117] "RemoveContainer" containerID="1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11" Oct 07 22:09:38 crc kubenswrapper[4871]: E1007 22:09:38.449685 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.473990 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.494308 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.512148 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.527136 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.527192 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.527208 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.527233 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.527249 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.531343 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.553541 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.572653 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.592146 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.609718 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.628002 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.630715 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.630844 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.630870 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.631111 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.631140 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.647484 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.668418 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.687269 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.708010 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.724388 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.735102 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.735262 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.735296 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.735326 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.735345 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.749404 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.796123 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206d3e5348638dc0d38e84033a65bb0e92df0c8107559813c3576dc1d3470209\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:20Z\\\",\\\"message\\\":\\\" (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444160 6345 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 22:09:20.444020 6345 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444091 6345 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:09:20.444147 6345 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1007 22:09:20.444972 6345 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:09:20.445050 6345 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:09:20.445066 6345 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:09:20.445094 6345 factory.go:656] Stopping watch factory\\\\nI1007 22:09:20.445127 6345 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22:09:20.445218 6345 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1007 22:09:20.445228 6345 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 22:09:20.445241 6345 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:09:20.445265 6345 handler.go:208] Removed *v1.Pod ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:38Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.839003 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.839069 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.839085 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.839107 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.839126 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.910610 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:38 crc kubenswrapper[4871]: E1007 22:09:38.910917 4871 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:38 crc kubenswrapper[4871]: E1007 22:09:38.911018 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs podName:684fd639-9051-4c1a-8bf0-9d1cb7b5cf59 nodeName:}" failed. No retries permitted until 2025-10-07 22:09:54.910992245 +0000 UTC m=+68.713690348 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs") pod "network-metrics-daemon-gbxdg" (UID: "684fd639-9051-4c1a-8bf0-9d1cb7b5cf59") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.942930 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.942991 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.943004 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.943025 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.943038 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:38Z","lastTransitionTime":"2025-10-07T22:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:38 crc kubenswrapper[4871]: I1007 22:09:38.982623 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:38 crc kubenswrapper[4871]: E1007 22:09:38.983067 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.046614 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.046719 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.046732 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.046758 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.046774 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.150386 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.150470 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.150486 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.150512 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.150529 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.253933 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.254011 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.254033 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.254066 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.254095 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.357833 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.357869 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.357880 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.357898 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.357910 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.455628 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/2.log" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.462320 4871 scope.go:117] "RemoveContainer" containerID="1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11" Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.462621 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.463001 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.463049 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.463065 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.463139 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.463160 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.486240 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.564049 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.566136 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.566191 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.566208 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.566602 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.566654 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.591427 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.608952 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.626522 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.631931 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.632090 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632170 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:10:11.632120257 +0000 UTC m=+85.434818330 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632207 4871 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632262 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:10:11.632244951 +0000 UTC m=+85.434943024 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.632263 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.632372 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632446 4871 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.632445 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632517 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:10:11.632503558 +0000 UTC m=+85.435201631 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632549 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632576 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632591 4871 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632613 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632664 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632684 4871 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632641 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 22:10:11.632628991 +0000 UTC m=+85.435327294 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.632766 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 22:10:11.632741504 +0000 UTC m=+85.435439587 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.647010 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.661685 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.669232 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.669260 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.669270 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.669286 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.669298 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.672998 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.686895 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.700165 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.717897 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.744837 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.767603 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.771896 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.771934 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.771948 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.771970 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.772041 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.792310 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.810698 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.829265 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:39Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.875514 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.875560 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.875571 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.875587 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.875597 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.979148 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.979226 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.979243 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.979270 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.979293 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:39Z","lastTransitionTime":"2025-10-07T22:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.981374 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.981481 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.981529 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:39 crc kubenswrapper[4871]: I1007 22:09:39.981380 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.981664 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:39 crc kubenswrapper[4871]: E1007 22:09:39.981707 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.083367 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.083411 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.083420 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.083437 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.083448 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:40Z","lastTransitionTime":"2025-10-07T22:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.187722 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.187857 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.187881 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.187946 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.187968 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:40Z","lastTransitionTime":"2025-10-07T22:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.292494 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.292552 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.292569 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.292596 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.292613 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:40Z","lastTransitionTime":"2025-10-07T22:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.395864 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.395939 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.395958 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.395988 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.396008 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:40Z","lastTransitionTime":"2025-10-07T22:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.499298 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.499341 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.499354 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.499373 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.499385 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:40Z","lastTransitionTime":"2025-10-07T22:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.602239 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.602567 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.602634 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.602707 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.602764 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:40Z","lastTransitionTime":"2025-10-07T22:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.706854 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.706925 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.706943 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.706973 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.707011 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:40Z","lastTransitionTime":"2025-10-07T22:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.810269 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.810564 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.810647 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.810765 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.810854 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:40Z","lastTransitionTime":"2025-10-07T22:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.914374 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.914869 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.914895 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.914923 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.914942 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:40Z","lastTransitionTime":"2025-10-07T22:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:40 crc kubenswrapper[4871]: I1007 22:09:40.982289 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:40 crc kubenswrapper[4871]: E1007 22:09:40.982752 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.018615 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.018694 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.018715 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.018742 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.018761 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.121722 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.122402 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.122507 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.122579 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.122640 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.226291 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.226577 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.226679 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.226744 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.226837 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.331226 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.331282 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.331301 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.331330 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.331349 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.435099 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.435166 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.435185 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.435219 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.435246 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.539844 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.539907 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.539928 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.539954 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.539972 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.612534 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.635849 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.642704 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.642757 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.642775 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.642831 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.642852 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.656171 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.682878 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.720019 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.745286 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.746288 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.746344 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.746358 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.746380 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.746393 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.766830 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.786772 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.804693 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.826761 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.850460 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.853649 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.853715 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.853755 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.853829 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.853860 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.876287 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.895692 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.914985 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.938432 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.957419 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.957500 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.957525 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.957561 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.957586 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:41Z","lastTransitionTime":"2025-10-07T22:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.964164 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.982243 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.982316 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:41 crc kubenswrapper[4871]: E1007 22:09:41.982477 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:41 crc kubenswrapper[4871]: E1007 22:09:41.982606 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.982911 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:41 crc kubenswrapper[4871]: E1007 22:09:41.983236 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:41 crc kubenswrapper[4871]: I1007 22:09:41.986115 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:41Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.045356 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.061451 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.061526 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.061554 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.061590 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.061618 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.062595 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.068570 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.090958 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.111059 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.131526 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.153879 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.164327 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.164397 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.164424 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.164457 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.164486 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.174748 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.190057 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.205358 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.219662 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.239103 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.259341 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.268147 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.268174 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.268186 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.268203 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.268217 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.278228 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.303151 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.318319 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.336050 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.360928 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:42Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.371762 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.371824 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.371841 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.371864 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.371880 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.474273 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.474355 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.474378 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.474407 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.474427 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.578637 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.579050 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.579197 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.579380 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.579509 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.683859 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.684356 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.684501 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.684648 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.684832 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.787854 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.787919 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.787938 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.787965 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.787986 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.890937 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.891012 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.891030 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.891056 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.891075 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.982399 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:42 crc kubenswrapper[4871]: E1007 22:09:42.982665 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.993358 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.993408 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.993419 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.993433 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:42 crc kubenswrapper[4871]: I1007 22:09:42.993442 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:42Z","lastTransitionTime":"2025-10-07T22:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.096427 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.096500 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.096517 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.096544 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.096563 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:43Z","lastTransitionTime":"2025-10-07T22:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.200179 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.200270 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.200299 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.200337 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.200366 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:43Z","lastTransitionTime":"2025-10-07T22:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.305370 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.305450 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.305465 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.305486 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.305501 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:43Z","lastTransitionTime":"2025-10-07T22:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.409149 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.409205 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.409216 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.409237 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.409250 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:43Z","lastTransitionTime":"2025-10-07T22:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.513040 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.513133 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.513153 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.513179 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.513198 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:43Z","lastTransitionTime":"2025-10-07T22:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.616558 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.616619 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.616638 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.616663 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.616681 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:43Z","lastTransitionTime":"2025-10-07T22:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.720435 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.720516 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.720536 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.720565 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.720585 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:43Z","lastTransitionTime":"2025-10-07T22:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.824863 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.824931 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.824950 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.824976 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.825003 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:43Z","lastTransitionTime":"2025-10-07T22:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.928662 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.928741 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.928759 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.928878 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.928922 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:43Z","lastTransitionTime":"2025-10-07T22:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.982120 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.982228 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:43 crc kubenswrapper[4871]: I1007 22:09:43.982325 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:43 crc kubenswrapper[4871]: E1007 22:09:43.982535 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:43 crc kubenswrapper[4871]: E1007 22:09:43.982705 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:43 crc kubenswrapper[4871]: E1007 22:09:43.982879 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.031323 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.031360 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.031372 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.031389 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.031404 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.134835 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.134940 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.134966 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.135751 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.136018 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.240214 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.240297 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.240320 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.240354 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.240377 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.344262 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.344330 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.344353 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.344381 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.344404 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.448000 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.448056 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.448072 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.448141 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.448167 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.508341 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.508422 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.508439 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.508466 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.508493 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: E1007 22:09:44.530628 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:44Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.536740 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.536831 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.536856 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.536882 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.536901 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: E1007 22:09:44.557006 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:44Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.562852 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.562932 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.562998 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.563026 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.563041 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: E1007 22:09:44.585264 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:44Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.590638 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.590714 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.590741 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.590775 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.590838 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: E1007 22:09:44.607159 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:44Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.613157 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.613209 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.613226 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.613247 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.613262 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: E1007 22:09:44.631683 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:44Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:44 crc kubenswrapper[4871]: E1007 22:09:44.631850 4871 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.634277 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.634311 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.634324 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.634347 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.634359 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.738111 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.738188 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.738209 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.738237 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.738259 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.841565 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.841622 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.841635 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.841658 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.841672 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.944879 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.944949 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.944966 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.944993 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.945015 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:44Z","lastTransitionTime":"2025-10-07T22:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:44 crc kubenswrapper[4871]: I1007 22:09:44.982218 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:44 crc kubenswrapper[4871]: E1007 22:09:44.982455 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.048843 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.048913 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.048935 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.048964 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.048986 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.151918 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.152003 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.152028 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.152055 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.152074 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.255276 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.255345 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.255368 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.255398 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.255421 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.359595 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.359650 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.359668 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.359694 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.359712 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.463381 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.463450 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.463468 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.463494 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.463512 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.567158 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.567231 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.567252 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.567278 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.567297 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.670976 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.671023 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.671039 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.671063 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.671082 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.775098 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.775178 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.775197 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.775222 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.775240 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.878513 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.878602 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.878621 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.878646 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.878667 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.981248 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.981386 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:45 crc kubenswrapper[4871]: E1007 22:09:45.981522 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:45 crc kubenswrapper[4871]: E1007 22:09:45.981595 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.982691 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.982856 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.982891 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.982901 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.982918 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:45 crc kubenswrapper[4871]: I1007 22:09:45.982930 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:45Z","lastTransitionTime":"2025-10-07T22:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:45 crc kubenswrapper[4871]: E1007 22:09:45.983276 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.087941 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.087982 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.087992 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.088008 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.088018 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:46Z","lastTransitionTime":"2025-10-07T22:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.192770 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.192842 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.192854 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.192874 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.192888 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:46Z","lastTransitionTime":"2025-10-07T22:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.296068 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.296130 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.296142 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.296164 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.296178 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:46Z","lastTransitionTime":"2025-10-07T22:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.399253 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.399320 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.399338 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.399365 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.399384 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:46Z","lastTransitionTime":"2025-10-07T22:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.502353 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.502423 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.502442 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.502472 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.502493 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:46Z","lastTransitionTime":"2025-10-07T22:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.606494 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.606563 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.606580 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.606607 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.606626 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:46Z","lastTransitionTime":"2025-10-07T22:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.710298 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.710387 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.710405 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.710437 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.710460 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:46Z","lastTransitionTime":"2025-10-07T22:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.813523 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.813604 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.813628 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.813663 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.813687 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:46Z","lastTransitionTime":"2025-10-07T22:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.916953 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.917020 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.917038 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.917071 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.917090 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:46Z","lastTransitionTime":"2025-10-07T22:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.982006 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:46 crc kubenswrapper[4871]: E1007 22:09:46.982228 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:46 crc kubenswrapper[4871]: I1007 22:09:46.999622 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:46Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.013138 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.026347 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.026415 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.026438 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.026473 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.026495 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.037109 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.058216 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.077491 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.096147 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.115085 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.130676 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.130723 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.130741 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.130770 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.130812 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.133435 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.154916 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.170686 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.187115 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.204994 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.225463 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.234366 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.234430 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.234454 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.234486 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.234508 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.253937 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.272328 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.289873 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.317050 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:47Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.337936 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.338013 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.338040 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.338075 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.338101 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.441544 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.441615 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.441634 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.441661 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.441681 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.545568 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.545631 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.545647 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.545670 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.545684 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.650156 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.650212 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.650230 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.650257 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.650275 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.753683 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.753724 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.753736 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.753755 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.753768 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.863342 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.863419 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.863456 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.863481 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.863493 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.966305 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.966360 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.966375 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.966393 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.966405 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:47Z","lastTransitionTime":"2025-10-07T22:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.981840 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.981901 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:47 crc kubenswrapper[4871]: E1007 22:09:47.981973 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:47 crc kubenswrapper[4871]: I1007 22:09:47.981901 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:47 crc kubenswrapper[4871]: E1007 22:09:47.982115 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:47 crc kubenswrapper[4871]: E1007 22:09:47.982131 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.069845 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.069916 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.069936 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.069964 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.069987 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:48Z","lastTransitionTime":"2025-10-07T22:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.173027 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.173100 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.173118 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.173142 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.173163 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:48Z","lastTransitionTime":"2025-10-07T22:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.276854 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.276907 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.276963 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.276992 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.277063 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:48Z","lastTransitionTime":"2025-10-07T22:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.380671 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.380746 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.380772 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.380840 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.380869 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:48Z","lastTransitionTime":"2025-10-07T22:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.488641 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.488694 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.488704 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.488721 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.488732 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:48Z","lastTransitionTime":"2025-10-07T22:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.593072 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.593130 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.593140 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.593163 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.593175 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:48Z","lastTransitionTime":"2025-10-07T22:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.697098 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.697167 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.697193 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.697226 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.697251 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:48Z","lastTransitionTime":"2025-10-07T22:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.800268 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.800362 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.800381 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.800410 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.800427 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:48Z","lastTransitionTime":"2025-10-07T22:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.903656 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.903756 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.903770 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.903817 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.903834 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:48Z","lastTransitionTime":"2025-10-07T22:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:48 crc kubenswrapper[4871]: I1007 22:09:48.982552 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:48 crc kubenswrapper[4871]: E1007 22:09:48.982725 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.007337 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.007423 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.007451 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.007487 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.007520 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.111719 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.111828 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.111856 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.111889 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.111915 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.215986 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.216040 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.216050 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.216076 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.216088 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.319332 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.319406 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.319424 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.319453 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.319475 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.423411 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.423492 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.423510 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.423538 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.423556 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.526410 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.526461 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.526470 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.526486 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.526497 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.630363 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.630438 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.630459 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.630488 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.630506 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.734141 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.734207 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.734226 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.734257 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.734276 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.837982 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.838056 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.838073 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.838099 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.838117 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.943698 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.943819 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.943840 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.943898 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.943919 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:49Z","lastTransitionTime":"2025-10-07T22:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.982375 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:49 crc kubenswrapper[4871]: E1007 22:09:49.982606 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.983025 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:49 crc kubenswrapper[4871]: I1007 22:09:49.983030 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:49 crc kubenswrapper[4871]: E1007 22:09:49.983270 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:49 crc kubenswrapper[4871]: E1007 22:09:49.983439 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.047933 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.048055 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.048080 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.048147 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.048169 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.152229 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.152330 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.152356 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.152392 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.152425 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.256293 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.256373 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.256391 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.256427 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.256541 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.360343 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.360410 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.360425 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.360449 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.360466 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.463894 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.463975 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.463994 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.464022 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.464042 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.567938 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.567988 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.568005 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.568030 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.568048 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.671926 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.671999 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.672020 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.672048 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.672073 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.776459 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.776535 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.776553 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.776579 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.776599 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.880083 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.880149 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.880167 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.880197 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.880217 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.981614 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:50 crc kubenswrapper[4871]: E1007 22:09:50.981959 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.983336 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.983377 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.983387 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.983406 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:50 crc kubenswrapper[4871]: I1007 22:09:50.983416 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:50Z","lastTransitionTime":"2025-10-07T22:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.088502 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.088565 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.088578 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.088601 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.088614 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:51Z","lastTransitionTime":"2025-10-07T22:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.191885 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.191957 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.191980 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.192014 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.192038 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:51Z","lastTransitionTime":"2025-10-07T22:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.295331 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.295389 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.295412 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.295437 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.295455 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:51Z","lastTransitionTime":"2025-10-07T22:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.398609 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.398686 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.398713 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.398754 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.398781 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:51Z","lastTransitionTime":"2025-10-07T22:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.502002 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.502048 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.502092 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.502121 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.502137 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:51Z","lastTransitionTime":"2025-10-07T22:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.605760 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.605858 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.605874 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.605898 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.605919 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:51Z","lastTransitionTime":"2025-10-07T22:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.709276 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.709317 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.709329 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.709348 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.709360 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:51Z","lastTransitionTime":"2025-10-07T22:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.812005 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.812077 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.812095 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.812124 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.812142 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:51Z","lastTransitionTime":"2025-10-07T22:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.916367 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.916441 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.916468 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.916498 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.916514 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:51Z","lastTransitionTime":"2025-10-07T22:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.981468 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.981487 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:51 crc kubenswrapper[4871]: E1007 22:09:51.981684 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:51 crc kubenswrapper[4871]: E1007 22:09:51.981870 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:51 crc kubenswrapper[4871]: I1007 22:09:51.981491 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:51 crc kubenswrapper[4871]: E1007 22:09:51.982013 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.020066 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.020108 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.020122 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.020141 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.020162 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.123478 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.123538 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.123555 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.123578 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.123591 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.226844 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.226884 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.226901 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.226926 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.226946 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.329607 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.329685 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.329707 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.329737 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.329758 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.445308 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.445387 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.445407 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.445433 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.445451 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.548378 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.548450 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.548469 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.548498 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.548516 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.651781 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.651908 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.651923 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.651947 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.651962 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.754350 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.754409 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.754420 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.754440 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.754453 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.856747 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.856838 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.856859 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.856883 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.856899 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.959640 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.959676 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.959689 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.959705 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.959719 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:52Z","lastTransitionTime":"2025-10-07T22:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:52 crc kubenswrapper[4871]: I1007 22:09:52.982259 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:52 crc kubenswrapper[4871]: E1007 22:09:52.982437 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.062462 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.062519 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.062531 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.062552 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.062568 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.165343 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.165408 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.165421 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.165443 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.165459 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.269022 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.269085 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.269097 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.269112 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.269123 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.371950 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.372044 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.372063 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.372092 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.372115 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.475088 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.475146 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.475162 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.475185 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.475201 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.578243 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.578360 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.578379 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.578405 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.578423 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.680924 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.680979 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.681011 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.681032 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.681049 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.784613 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.784670 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.784687 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.784715 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.784733 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.888970 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.889038 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.889054 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.889085 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.889105 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.981480 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.981542 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.981680 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:53 crc kubenswrapper[4871]: E1007 22:09:53.981833 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:53 crc kubenswrapper[4871]: E1007 22:09:53.982587 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:53 crc kubenswrapper[4871]: E1007 22:09:53.982761 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.983235 4871 scope.go:117] "RemoveContainer" containerID="1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11" Oct 07 22:09:53 crc kubenswrapper[4871]: E1007 22:09:53.983639 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.992537 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.992613 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.992641 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.992671 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:53 crc kubenswrapper[4871]: I1007 22:09:53.992695 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:53Z","lastTransitionTime":"2025-10-07T22:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.095990 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.096051 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.096064 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.096086 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.096102 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.199061 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.199102 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.199112 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.199129 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.199141 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.301864 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.301909 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.301923 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.301945 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.301956 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.404723 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.404781 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.404811 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.404831 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.404844 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.507889 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.507972 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.507992 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.508022 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.508051 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.611302 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.611361 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.611373 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.611392 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.611406 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.714611 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.714690 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.714717 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.714755 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.714779 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.817999 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.818053 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.818067 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.818087 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.818098 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.921578 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.921668 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.921692 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.921742 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.921763 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.923815 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:54 crc kubenswrapper[4871]: E1007 22:09:54.924040 4871 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:54 crc kubenswrapper[4871]: E1007 22:09:54.924131 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs podName:684fd639-9051-4c1a-8bf0-9d1cb7b5cf59 nodeName:}" failed. No retries permitted until 2025-10-07 22:10:26.924089512 +0000 UTC m=+100.726787625 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs") pod "network-metrics-daemon-gbxdg" (UID: "684fd639-9051-4c1a-8bf0-9d1cb7b5cf59") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.981923 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:54 crc kubenswrapper[4871]: E1007 22:09:54.982142 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.984723 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.984824 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.984839 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.984855 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:54 crc kubenswrapper[4871]: I1007 22:09:54.984869 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:54Z","lastTransitionTime":"2025-10-07T22:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:54 crc kubenswrapper[4871]: E1007 22:09:54.998955 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:54Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.002954 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.002993 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.003004 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.003021 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.003035 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: E1007 22:09:55.020838 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:55Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.027465 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.027524 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.027547 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.027580 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.027607 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: E1007 22:09:55.041557 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:55Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.046344 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.046464 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.046478 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.046496 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.046512 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: E1007 22:09:55.059148 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:55Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.063758 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.063842 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.063858 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.063877 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.063890 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: E1007 22:09:55.080905 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:55Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:55 crc kubenswrapper[4871]: E1007 22:09:55.081043 4871 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.082715 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.082749 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.082759 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.082773 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.082784 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.186230 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.186322 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.186349 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.186384 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.186410 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.290116 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.290169 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.290178 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.290197 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.290210 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.394191 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.394271 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.394292 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.394359 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.394380 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.497846 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.497893 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.497905 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.497927 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.497944 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.601406 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.601456 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.601465 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.601483 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.601494 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.705388 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.705427 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.705438 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.705456 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.705470 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.808545 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.808614 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.808641 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.808674 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.808699 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.911429 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.911504 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.911528 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.911599 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.911626 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:55Z","lastTransitionTime":"2025-10-07T22:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.982327 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.982509 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:55 crc kubenswrapper[4871]: E1007 22:09:55.982643 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:55 crc kubenswrapper[4871]: I1007 22:09:55.982672 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:55 crc kubenswrapper[4871]: E1007 22:09:55.982889 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:55 crc kubenswrapper[4871]: E1007 22:09:55.982968 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.014959 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.015018 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.015031 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.015053 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.015073 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.118099 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.118147 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.118162 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.118181 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.118194 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.221366 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.221424 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.221434 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.221454 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.221465 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.324239 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.324324 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.324343 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.324373 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.324398 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.428073 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.428136 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.428147 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.428164 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.428175 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.526408 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nv4jc_53708429-59b4-4319-bdb7-8a922a551e59/kube-multus/0.log" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.526496 4871 generic.go:334] "Generic (PLEG): container finished" podID="53708429-59b4-4319-bdb7-8a922a551e59" containerID="ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674" exitCode=1 Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.526547 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nv4jc" event={"ID":"53708429-59b4-4319-bdb7-8a922a551e59","Type":"ContainerDied","Data":"ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.527771 4871 scope.go:117] "RemoveContainer" containerID="ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.539642 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.539919 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.540085 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.540234 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.540352 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.571027 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.588259 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.607709 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.629752 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.643848 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.643897 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.643909 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.643928 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.643940 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.646973 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.665525 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.682690 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.697056 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.711312 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.722988 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.736364 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.747386 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.747450 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.747467 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.747493 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.747508 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.752272 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.766159 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.781303 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"2025-10-07T22:09:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56\\\\n2025-10-07T22:09:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56 to /host/opt/cni/bin/\\\\n2025-10-07T22:09:10Z [verbose] multus-daemon started\\\\n2025-10-07T22:09:10Z [verbose] Readiness Indicator file check\\\\n2025-10-07T22:09:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.792022 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.806090 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.819765 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.850152 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.850189 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.850204 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.850223 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.850240 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.953060 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.953140 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.953156 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.953173 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.953185 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:56Z","lastTransitionTime":"2025-10-07T22:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.982014 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:56 crc kubenswrapper[4871]: E1007 22:09:56.982230 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:56 crc kubenswrapper[4871]: I1007 22:09:56.996055 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.000398 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:56Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.017963 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.041874 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.055728 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.055836 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.055860 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.055893 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.055919 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.065893 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.087333 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.112569 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.127450 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.142375 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.156549 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.159678 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.159718 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.159728 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.159743 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.159754 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.171343 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.183747 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.199977 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.215447 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.239377 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.262634 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.262730 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.262750 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.262779 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.262824 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.272227 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.293167 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"2025-10-07T22:09:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56\\\\n2025-10-07T22:09:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56 to /host/opt/cni/bin/\\\\n2025-10-07T22:09:10Z [verbose] multus-daemon started\\\\n2025-10-07T22:09:10Z [verbose] Readiness Indicator file check\\\\n2025-10-07T22:09:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.339148 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.365039 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.365108 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.365128 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.365158 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.365176 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.468215 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.468288 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.468316 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.468344 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.468363 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.533845 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nv4jc_53708429-59b4-4319-bdb7-8a922a551e59/kube-multus/0.log" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.533994 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nv4jc" event={"ID":"53708429-59b4-4319-bdb7-8a922a551e59","Type":"ContainerStarted","Data":"757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.556040 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.570859 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.570893 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.570905 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.570923 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.570937 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.575439 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.598052 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"2025-10-07T22:09:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56\\\\n2025-10-07T22:09:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56 to /host/opt/cni/bin/\\\\n2025-10-07T22:09:10Z [verbose] multus-daemon started\\\\n2025-10-07T22:09:10Z [verbose] Readiness Indicator file check\\\\n2025-10-07T22:09:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.615492 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.635689 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.652715 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.673852 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.673895 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.673915 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.673957 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.673977 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.677683 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.709911 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.729742 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcfed2fc-23c7-4293-9d0f-b1fe9ac20394\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62ddf0800fa21c83bb765ca4ae13168e4ba543ae0b9ae2c52a264b019b17561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.754040 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.773925 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.778052 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.778107 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.778122 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.778142 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.778156 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.793054 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.811970 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.831567 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.855736 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.878241 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.882674 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.882727 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.882745 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.882772 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.882820 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.900714 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.940143 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:09:57Z is after 2025-08-24T17:21:41Z" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.981992 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:57 crc kubenswrapper[4871]: E1007 22:09:57.982196 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.982545 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:57 crc kubenswrapper[4871]: E1007 22:09:57.982644 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.982969 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:57 crc kubenswrapper[4871]: E1007 22:09:57.983162 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.985665 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.986063 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.986093 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.986121 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:57 crc kubenswrapper[4871]: I1007 22:09:57.986133 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:57Z","lastTransitionTime":"2025-10-07T22:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.090939 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.090999 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.091018 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.091049 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.091068 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:58Z","lastTransitionTime":"2025-10-07T22:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.194614 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.194669 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.194683 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.194706 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.194722 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:58Z","lastTransitionTime":"2025-10-07T22:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.297749 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.297837 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.297855 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.297881 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.297903 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:58Z","lastTransitionTime":"2025-10-07T22:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.400713 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.400781 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.400822 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.400844 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.400855 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:58Z","lastTransitionTime":"2025-10-07T22:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.503451 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.503494 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.503505 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.503520 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.503532 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:58Z","lastTransitionTime":"2025-10-07T22:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.606873 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.606929 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.606947 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.606975 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.606992 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:58Z","lastTransitionTime":"2025-10-07T22:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.710499 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.710566 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.710587 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.710613 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.710635 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:58Z","lastTransitionTime":"2025-10-07T22:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.813234 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.813290 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.813308 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.813335 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.813351 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:58Z","lastTransitionTime":"2025-10-07T22:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.916285 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.916333 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.916348 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.916368 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.916380 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:58Z","lastTransitionTime":"2025-10-07T22:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:58 crc kubenswrapper[4871]: I1007 22:09:58.981556 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:09:58 crc kubenswrapper[4871]: E1007 22:09:58.982002 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.019480 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.019537 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.019550 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.019571 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.019586 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.123007 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.123075 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.123089 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.123115 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.123129 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.225836 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.225905 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.225926 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.225955 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.225975 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.329033 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.329112 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.329131 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.329160 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.329179 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.433136 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.433200 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.433220 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.433248 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.433267 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.537557 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.537614 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.537639 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.537677 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.537704 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.642954 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.643044 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.643064 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.643119 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.643138 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.747847 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.747920 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.747938 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.747972 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.747992 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.851788 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.851866 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.851881 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.851903 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.851920 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.955734 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.955848 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.955875 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.955904 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.955921 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:09:59Z","lastTransitionTime":"2025-10-07T22:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.982100 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.982125 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:09:59 crc kubenswrapper[4871]: I1007 22:09:59.982175 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:09:59 crc kubenswrapper[4871]: E1007 22:09:59.982288 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:09:59 crc kubenswrapper[4871]: E1007 22:09:59.982466 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:09:59 crc kubenswrapper[4871]: E1007 22:09:59.982705 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.060237 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.060313 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.060331 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.060360 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.060379 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.163568 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.163619 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.163637 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.163663 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.163682 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.266908 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.266979 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.266996 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.267441 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.267502 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.370733 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.370831 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.370850 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.370878 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.370897 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.474559 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.474629 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.474647 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.474673 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.474694 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.578314 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.578381 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.578398 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.578424 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.578442 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.682614 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.682659 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.682676 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.682694 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.682706 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.785599 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.785657 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.785670 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.785692 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.785705 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.889876 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.889948 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.889974 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.890009 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.890030 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.981973 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:00 crc kubenswrapper[4871]: E1007 22:10:00.982205 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.992833 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.992906 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.992930 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.992966 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:00 crc kubenswrapper[4871]: I1007 22:10:00.992990 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:00Z","lastTransitionTime":"2025-10-07T22:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.097140 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.097201 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.097225 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.097257 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.097284 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:01Z","lastTransitionTime":"2025-10-07T22:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.201153 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.201243 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.201268 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.201305 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.201328 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:01Z","lastTransitionTime":"2025-10-07T22:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.310857 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.310939 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.310961 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.310994 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.311019 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:01Z","lastTransitionTime":"2025-10-07T22:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.413843 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.413897 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.413911 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.413932 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.413949 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:01Z","lastTransitionTime":"2025-10-07T22:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.517962 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.518047 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.518074 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.518106 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.518126 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:01Z","lastTransitionTime":"2025-10-07T22:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.621413 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.621875 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.621908 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.621936 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.621955 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:01Z","lastTransitionTime":"2025-10-07T22:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.725653 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.725747 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.725774 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.725843 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.725873 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:01Z","lastTransitionTime":"2025-10-07T22:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.828862 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.828965 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.828984 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.829010 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.829028 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:01Z","lastTransitionTime":"2025-10-07T22:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.932415 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.932484 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.932504 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.932532 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.932551 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:01Z","lastTransitionTime":"2025-10-07T22:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.982088 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.982208 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:01 crc kubenswrapper[4871]: E1007 22:10:01.982298 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:01 crc kubenswrapper[4871]: I1007 22:10:01.982103 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:01 crc kubenswrapper[4871]: E1007 22:10:01.982441 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:01 crc kubenswrapper[4871]: E1007 22:10:01.983452 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.036159 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.036221 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.036239 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.036263 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.036281 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.139029 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.139096 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.139114 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.139143 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.139166 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.242286 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.242347 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.242372 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.242402 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.242426 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.345247 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.345419 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.345442 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.345466 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.345484 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.448761 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.448859 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.448883 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.448910 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.448927 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.551270 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.551344 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.551368 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.551433 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.551460 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.654691 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.654753 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.654773 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.654836 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.654864 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.757349 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.757400 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.757418 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.757448 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.757466 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.861617 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.861682 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.861700 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.861727 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.861747 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.965449 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.965524 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.965551 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.965590 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.965618 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:02Z","lastTransitionTime":"2025-10-07T22:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:02 crc kubenswrapper[4871]: I1007 22:10:02.981885 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:02 crc kubenswrapper[4871]: E1007 22:10:02.982185 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.070389 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.070450 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.070471 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.070496 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.070517 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:03Z","lastTransitionTime":"2025-10-07T22:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.173882 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.173936 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.173955 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.173977 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.174002 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:03Z","lastTransitionTime":"2025-10-07T22:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.277655 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.277718 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.277735 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.277781 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.277855 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:03Z","lastTransitionTime":"2025-10-07T22:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.381873 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.381942 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.381960 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.381989 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.382008 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:03Z","lastTransitionTime":"2025-10-07T22:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.486183 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.486268 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.486295 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.486330 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.486352 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:03Z","lastTransitionTime":"2025-10-07T22:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.589253 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.589625 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.589708 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.589812 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.589904 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:03Z","lastTransitionTime":"2025-10-07T22:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.693954 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.694378 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.694648 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.694945 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.695158 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:03Z","lastTransitionTime":"2025-10-07T22:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.798443 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.798511 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.798534 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.798565 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.798585 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:03Z","lastTransitionTime":"2025-10-07T22:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.902123 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.902221 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.902241 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.902269 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.902317 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:03Z","lastTransitionTime":"2025-10-07T22:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.981327 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.981356 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:03 crc kubenswrapper[4871]: I1007 22:10:03.981326 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:03 crc kubenswrapper[4871]: E1007 22:10:03.981498 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:03 crc kubenswrapper[4871]: E1007 22:10:03.981614 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:03 crc kubenswrapper[4871]: E1007 22:10:03.981744 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.005944 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.006006 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.006030 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.006062 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.006088 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.109516 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.110008 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.110174 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.110338 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.110535 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.214199 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.214647 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.214855 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.215027 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.215205 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.318756 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.318851 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.318869 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.318900 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.318919 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.423308 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.423696 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.423844 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.423975 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.424075 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.528100 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.528716 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.528911 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.529055 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.529169 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.633445 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.633510 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.633528 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.633555 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.633779 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.738121 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.738203 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.738228 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.738263 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.738287 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.841623 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.841678 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.841690 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.841711 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.841726 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.945633 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.945740 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.946222 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.946295 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.946621 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:04Z","lastTransitionTime":"2025-10-07T22:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:04 crc kubenswrapper[4871]: I1007 22:10:04.982422 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:04 crc kubenswrapper[4871]: E1007 22:10:04.982715 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.049940 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.050023 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.050042 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.050075 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.050094 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.154125 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.154186 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.154204 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.154233 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.154253 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.219329 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.219420 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.219437 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.219466 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.219482 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: E1007 22:10:05.240447 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:05Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.245403 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.245504 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.245535 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.245570 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.245597 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: E1007 22:10:05.270512 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:05Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.276736 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.276828 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.276846 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.276872 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.276892 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: E1007 22:10:05.296256 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:05Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.300949 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.301027 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.301052 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.301084 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.301112 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: E1007 22:10:05.323840 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:05Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.329392 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.329450 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.329462 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.329480 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.329494 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: E1007 22:10:05.348510 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:05Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:05 crc kubenswrapper[4871]: E1007 22:10:05.348737 4871 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.351516 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.351586 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.351607 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.351635 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.351656 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.455479 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.455547 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.455563 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.455590 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.455608 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.559559 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.559674 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.559695 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.559724 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.559751 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.663048 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.663107 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.663125 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.663152 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.663175 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.766158 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.766227 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.766252 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.766286 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.766307 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.869142 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.869176 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.869185 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.869203 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.869213 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.972646 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.972722 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.972741 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.972772 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.972821 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:05Z","lastTransitionTime":"2025-10-07T22:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.982063 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.982095 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:05 crc kubenswrapper[4871]: I1007 22:10:05.982139 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:05 crc kubenswrapper[4871]: E1007 22:10:05.982227 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:05 crc kubenswrapper[4871]: E1007 22:10:05.982662 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:05 crc kubenswrapper[4871]: E1007 22:10:05.982891 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.076869 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.076918 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.076931 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.076953 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.076968 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:06Z","lastTransitionTime":"2025-10-07T22:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.180057 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.180103 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.180112 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.180128 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.180137 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:06Z","lastTransitionTime":"2025-10-07T22:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.284161 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.284246 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.284273 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.284306 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.284331 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:06Z","lastTransitionTime":"2025-10-07T22:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.387933 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.388010 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.388036 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.388064 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.388085 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:06Z","lastTransitionTime":"2025-10-07T22:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.491362 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.491441 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.491466 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.491500 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.491525 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:06Z","lastTransitionTime":"2025-10-07T22:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.595426 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.595495 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.595513 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.595540 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.595559 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:06Z","lastTransitionTime":"2025-10-07T22:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.699481 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.699543 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.699561 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.699591 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.699610 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:06Z","lastTransitionTime":"2025-10-07T22:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.803204 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.803265 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.803283 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.803310 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.803328 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:06Z","lastTransitionTime":"2025-10-07T22:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.906786 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.906886 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.906906 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.906937 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.906958 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:06Z","lastTransitionTime":"2025-10-07T22:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.981993 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:06 crc kubenswrapper[4871]: E1007 22:10:06.982218 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:06 crc kubenswrapper[4871]: I1007 22:10:06.983247 4871 scope.go:117] "RemoveContainer" containerID="1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.006628 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.011870 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.011927 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.011944 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.011976 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.011993 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.024243 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.048767 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.081733 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.101722 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcfed2fc-23c7-4293-9d0f-b1fe9ac20394\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62ddf0800fa21c83bb765ca4ae13168e4ba543ae0b9ae2c52a264b019b17561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.115092 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.115634 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.115929 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.116383 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.117045 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.125207 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.146276 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.165298 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.183173 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.200486 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.221522 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.221567 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.221586 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.221611 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.221628 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.222516 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.248263 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.267093 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.283109 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.299280 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.317412 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.324981 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.325031 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.325051 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.325077 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.325098 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.341382 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"2025-10-07T22:09:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56\\\\n2025-10-07T22:09:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56 to /host/opt/cni/bin/\\\\n2025-10-07T22:09:10Z [verbose] multus-daemon started\\\\n2025-10-07T22:09:10Z [verbose] Readiness Indicator file check\\\\n2025-10-07T22:09:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.353568 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.429266 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.429326 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.429340 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.429362 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.429375 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.532916 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.532995 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.533014 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.533043 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.533066 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.572904 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/2.log" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.576022 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.576611 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.591313 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcfed2fc-23c7-4293-9d0f-b1fe9ac20394\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62ddf0800fa21c83bb765ca4ae13168e4ba543ae0b9ae2c52a264b019b17561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.606604 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.621733 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.633649 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.643750 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.643813 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.643827 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.643871 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.643885 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.649646 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.664353 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.685281 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.705268 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.722225 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.734848 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.746808 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.746847 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.746857 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.746877 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.746889 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.751749 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.770845 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.789335 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"2025-10-07T22:09:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56\\\\n2025-10-07T22:09:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56 to /host/opt/cni/bin/\\\\n2025-10-07T22:09:10Z [verbose] multus-daemon started\\\\n2025-10-07T22:09:10Z [verbose] Readiness Indicator file check\\\\n2025-10-07T22:09:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.801912 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.823954 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.841255 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.849166 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.849200 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.849210 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.849227 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.849238 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.866628 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.900329 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:07Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.971361 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.971445 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.971470 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.971925 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.972311 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:07Z","lastTransitionTime":"2025-10-07T22:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.981475 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.981501 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:07 crc kubenswrapper[4871]: I1007 22:10:07.981650 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:07 crc kubenswrapper[4871]: E1007 22:10:07.981671 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:07 crc kubenswrapper[4871]: E1007 22:10:07.981901 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:07 crc kubenswrapper[4871]: E1007 22:10:07.981984 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.075464 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.075534 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.075543 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.075558 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.075569 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:08Z","lastTransitionTime":"2025-10-07T22:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.179170 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.179220 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.179237 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.179261 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.179282 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:08Z","lastTransitionTime":"2025-10-07T22:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.282992 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.283363 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.283551 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.283715 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.283920 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:08Z","lastTransitionTime":"2025-10-07T22:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.387095 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.387171 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.387192 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.387220 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.387238 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:08Z","lastTransitionTime":"2025-10-07T22:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.491296 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.491365 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.491388 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.491417 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.491437 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:08Z","lastTransitionTime":"2025-10-07T22:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.583366 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/3.log" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.584617 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/2.log" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.589609 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" exitCode=1 Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.589685 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.589759 4871 scope.go:117] "RemoveContainer" containerID="1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.591104 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:10:08 crc kubenswrapper[4871]: E1007 22:10:08.591379 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.594402 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.594476 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.594495 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.594527 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.594548 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:08Z","lastTransitionTime":"2025-10-07T22:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.615612 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.637365 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.659942 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"2025-10-07T22:09:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56\\\\n2025-10-07T22:09:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56 to /host/opt/cni/bin/\\\\n2025-10-07T22:09:10Z [verbose] multus-daemon started\\\\n2025-10-07T22:09:10Z [verbose] Readiness Indicator file check\\\\n2025-10-07T22:09:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.677201 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.699389 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.699462 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.699480 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.699507 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.699527 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:08Z","lastTransitionTime":"2025-10-07T22:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.701686 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.721245 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.745900 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.778955 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a8c84ddb70d18ea62a5ed0ba3614a059ff2cc19bb230a32db18bfdda484df11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:38Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}\\\\nI1007 22:09:38.103184 6585 services_controller.go:360] Finished syncing service ingress-canary on namespace openshift-ingress-canary for network=default : 9.850552ms\\\\nI1007 22:09:38.103216 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1007 22:09:38.103241 6585 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 10.871109ms\\\\nI1007 22:09:38.103266 6585 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}\\\\nI1007 22:09:38.103291 6585 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-storage-version-migrator-operator for network=default : 9.266696ms\\\\nI1007 22:09:38.103560 6585 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:09:38.103621 6585 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 22:09:38.103713 6585 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:10:08Z\\\",\\\"message\\\":\\\"rom k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.096778 6974 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.096997 6974 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.097343 6974 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.097617 6974 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.098030 6974 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:10:08.098085 6974 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 22:10:08.098102 6974 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:10:08.098107 6974 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:10:08.098157 6974 factory.go:656] Stopping watch factory\\\\nI1007 22:10:08.098178 6974 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:10:08.098209 6974 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:10:08.098217 6974 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 22:10:08.098224 6974 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.797878 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcfed2fc-23c7-4293-9d0f-b1fe9ac20394\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62ddf0800fa21c83bb765ca4ae13168e4ba543ae0b9ae2c52a264b019b17561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.803160 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.803233 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.803252 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.803286 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.803308 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:08Z","lastTransitionTime":"2025-10-07T22:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.822879 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.845765 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.869515 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.888036 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.907476 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.907577 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.907597 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.907625 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.907648 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:08Z","lastTransitionTime":"2025-10-07T22:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.915376 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.936899 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.959448 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.979040 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.981722 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:08 crc kubenswrapper[4871]: E1007 22:10:08.981958 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:08 crc kubenswrapper[4871]: I1007 22:10:08.997382 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:08Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.011146 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.011220 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.011243 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.011272 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.011292 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.115074 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.115144 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.115162 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.115195 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.115212 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.219765 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.219869 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.219891 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.219919 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.219940 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.322881 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.322970 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.322991 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.323019 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.323039 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.426639 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.426696 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.426713 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.426742 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.426760 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.529594 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.529646 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.529664 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.529688 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.529708 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.596226 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/3.log" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.601646 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:10:09 crc kubenswrapper[4871]: E1007 22:10:09.601827 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.625008 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.632002 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.632065 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.632085 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.632112 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.632132 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.645433 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.666055 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"2025-10-07T22:09:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56\\\\n2025-10-07T22:09:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56 to /host/opt/cni/bin/\\\\n2025-10-07T22:09:10Z [verbose] multus-daemon started\\\\n2025-10-07T22:09:10Z [verbose] Readiness Indicator file check\\\\n2025-10-07T22:09:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.683006 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.700331 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.725753 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.735650 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.735695 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.735709 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.735731 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.735745 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.764047 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:10:08Z\\\",\\\"message\\\":\\\"rom k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.096778 6974 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.096997 6974 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.097343 6974 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.097617 6974 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.098030 6974 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:10:08.098085 6974 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 22:10:08.098102 6974 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:10:08.098107 6974 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:10:08.098157 6974 factory.go:656] Stopping watch factory\\\\nI1007 22:10:08.098178 6974 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:10:08.098209 6974 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:10:08.098217 6974 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 22:10:08.098224 6974 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:10:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.785474 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.808980 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.831368 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.843627 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.843658 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.843669 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.843687 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.843700 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.850737 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.869328 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.885932 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcfed2fc-23c7-4293-9d0f-b1fe9ac20394\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62ddf0800fa21c83bb765ca4ae13168e4ba543ae0b9ae2c52a264b019b17561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.906950 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.932488 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.946534 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.946573 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.946592 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.946617 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.946635 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:09Z","lastTransitionTime":"2025-10-07T22:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.953163 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.971990 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.981855 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.981894 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.981940 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:09 crc kubenswrapper[4871]: E1007 22:10:09.982018 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:09 crc kubenswrapper[4871]: E1007 22:10:09.982193 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:09 crc kubenswrapper[4871]: E1007 22:10:09.982249 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:09 crc kubenswrapper[4871]: I1007 22:10:09.991690 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:09Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.057672 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.057746 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.057768 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.057822 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.057851 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.162710 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.162882 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.162908 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.162937 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.162959 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.267186 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.267251 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.267271 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.267297 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.267314 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.372114 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.372192 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.372210 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.372239 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.372257 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.475634 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.475700 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.475718 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.475741 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.475758 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.579312 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.579366 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.579382 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.579406 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.579422 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.683386 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.683443 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.683466 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.683495 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.683520 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.787277 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.787358 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.787392 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.787422 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.787442 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.891743 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.891828 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.891844 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.891867 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.891884 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.982002 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:10 crc kubenswrapper[4871]: E1007 22:10:10.982201 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.995113 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.995184 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.995204 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.995230 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:10 crc kubenswrapper[4871]: I1007 22:10:10.995248 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:10Z","lastTransitionTime":"2025-10-07T22:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.099469 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.099536 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.099561 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.099593 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.099614 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:11Z","lastTransitionTime":"2025-10-07T22:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.203820 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.203904 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.203923 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.203951 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.203971 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:11Z","lastTransitionTime":"2025-10-07T22:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.307868 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.307934 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.307952 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.307978 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.307998 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:11Z","lastTransitionTime":"2025-10-07T22:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.412029 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.412096 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.412116 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.412143 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.412165 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:11Z","lastTransitionTime":"2025-10-07T22:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.516072 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.516140 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.516161 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.516190 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.516210 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:11Z","lastTransitionTime":"2025-10-07T22:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.619962 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.620036 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.620058 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.620086 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.620108 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:11Z","lastTransitionTime":"2025-10-07T22:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.648496 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.648717 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.648775 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.648888 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.648787449 +0000 UTC m=+149.451485592 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.648984 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649007 4871 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649018 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649057 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649077 4871 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649089 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.649066466 +0000 UTC m=+149.451764579 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.649125 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649151 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.649124368 +0000 UTC m=+149.451822471 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649206 4871 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649266 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.649252541 +0000 UTC m=+149.451950654 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649343 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649413 4871 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649438 4871 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.649547 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.649516938 +0000 UTC m=+149.452215051 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.724355 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.724435 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.724455 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.724488 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.724559 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:11Z","lastTransitionTime":"2025-10-07T22:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.828501 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.828567 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.828584 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.828610 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.828629 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:11Z","lastTransitionTime":"2025-10-07T22:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.932133 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.932206 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.932222 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.932246 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.932262 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:11Z","lastTransitionTime":"2025-10-07T22:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.981884 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.981935 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:11 crc kubenswrapper[4871]: I1007 22:10:11.981891 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.982067 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.982251 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:11 crc kubenswrapper[4871]: E1007 22:10:11.982320 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.036094 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.036178 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.036194 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.036213 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.036226 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.140395 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.140452 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.140463 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.140484 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.140543 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.244370 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.244438 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.244457 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.244491 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.244511 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.348623 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.349149 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.349205 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.349233 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.349252 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.454363 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.454447 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.454470 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.454502 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.454533 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.559467 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.559546 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.559568 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.559599 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.559623 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.663008 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.663059 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.663078 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.663104 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.663124 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.766115 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.766185 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.766205 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.766234 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.766254 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.868489 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.868546 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.868561 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.868584 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.868600 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.971986 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.972050 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.972067 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.972097 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.972116 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:12Z","lastTransitionTime":"2025-10-07T22:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:12 crc kubenswrapper[4871]: I1007 22:10:12.981363 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:12 crc kubenswrapper[4871]: E1007 22:10:12.981536 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.075730 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.075854 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.075876 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.075904 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.075924 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:13Z","lastTransitionTime":"2025-10-07T22:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.179379 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.179438 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.179456 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.179485 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.179503 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:13Z","lastTransitionTime":"2025-10-07T22:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.283289 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.283357 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.283375 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.283404 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.283422 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:13Z","lastTransitionTime":"2025-10-07T22:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.387449 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.387530 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.387546 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.387571 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.387589 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:13Z","lastTransitionTime":"2025-10-07T22:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.491505 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.491609 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.491638 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.491674 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.491701 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:13Z","lastTransitionTime":"2025-10-07T22:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.595490 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.595557 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.595571 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.595593 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.595622 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:13Z","lastTransitionTime":"2025-10-07T22:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.699075 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.699149 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.699168 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.699201 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.699221 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:13Z","lastTransitionTime":"2025-10-07T22:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.803229 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.803315 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.803336 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.803367 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.803386 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:13Z","lastTransitionTime":"2025-10-07T22:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.906505 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.906594 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.906619 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.906656 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.906684 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:13Z","lastTransitionTime":"2025-10-07T22:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.982371 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.982427 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:13 crc kubenswrapper[4871]: I1007 22:10:13.982444 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:13 crc kubenswrapper[4871]: E1007 22:10:13.982629 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:13 crc kubenswrapper[4871]: E1007 22:10:13.982786 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:13 crc kubenswrapper[4871]: E1007 22:10:13.982967 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.010360 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.010452 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.010476 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.010513 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.010536 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.114091 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.114159 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.114177 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.114202 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.114220 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.217907 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.217994 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.218014 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.218042 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.218063 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.321787 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.321889 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.321907 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.321945 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.321967 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.425493 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.425566 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.425584 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.425611 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.425629 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.529632 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.529703 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.529725 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.529760 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.529783 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.633004 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.633065 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.633082 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.633106 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.633124 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.737213 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.737303 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.737348 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.737376 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.737395 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.840399 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.840488 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.840510 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.840544 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.840568 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.949615 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.950054 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.950075 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.950104 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.950123 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:14Z","lastTransitionTime":"2025-10-07T22:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:14 crc kubenswrapper[4871]: I1007 22:10:14.982413 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:14 crc kubenswrapper[4871]: E1007 22:10:14.982686 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.053866 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.053953 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.053975 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.054010 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.054037 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.158065 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.158204 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.158229 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.158267 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.158292 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.262901 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.262989 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.263007 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.263039 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.263060 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.366605 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.366671 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.366690 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.366717 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.366737 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.471324 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.471393 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.471409 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.471438 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.471458 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.506622 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.506707 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.506737 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.506775 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.506834 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: E1007 22:10:15.528968 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.535430 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.535493 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.535512 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.535539 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.535613 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: E1007 22:10:15.558376 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.562706 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.562740 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.562752 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.562776 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.562809 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: E1007 22:10:15.577784 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.583476 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.583569 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.583592 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.583622 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.583642 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: E1007 22:10:15.602655 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.608551 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.608600 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.608612 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.608637 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.608651 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: E1007 22:10:15.623314 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:15Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:15 crc kubenswrapper[4871]: E1007 22:10:15.623460 4871 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.625289 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.625351 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.625375 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.625410 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.625435 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.728752 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.728859 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.728885 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.728916 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.728940 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.832186 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.832253 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.832281 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.832310 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.832335 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.936316 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.936394 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.936411 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.936440 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.936461 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:15Z","lastTransitionTime":"2025-10-07T22:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.981370 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.981457 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:15 crc kubenswrapper[4871]: I1007 22:10:15.981369 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:15 crc kubenswrapper[4871]: E1007 22:10:15.981597 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:15 crc kubenswrapper[4871]: E1007 22:10:15.981785 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:15 crc kubenswrapper[4871]: E1007 22:10:15.982006 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.040648 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.040705 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.040723 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.040750 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.040769 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.144094 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.144153 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.144170 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.144196 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.144224 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.247827 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.248064 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.248087 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.248151 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.248172 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.351971 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.352058 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.352082 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.352115 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.352138 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.456205 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.456262 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.456291 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.456313 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.456331 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.560308 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.560365 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.560384 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.560413 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.560431 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.664033 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.664101 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.664120 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.664144 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.664162 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.767511 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.767560 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.767576 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.767601 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.767619 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.871202 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.871257 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.871274 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.871298 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.871315 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.974364 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.974434 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.974456 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.974484 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.974506 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:16Z","lastTransitionTime":"2025-10-07T22:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:16 crc kubenswrapper[4871]: I1007 22:10:16.981884 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:16 crc kubenswrapper[4871]: E1007 22:10:16.982085 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.003616 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.020879 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.039961 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.061612 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.078039 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.078111 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.078133 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.078162 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.078182 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:17Z","lastTransitionTime":"2025-10-07T22:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.081981 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.107045 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"2025-10-07T22:09:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56\\\\n2025-10-07T22:09:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56 to /host/opt/cni/bin/\\\\n2025-10-07T22:09:10Z [verbose] multus-daemon started\\\\n2025-10-07T22:09:10Z [verbose] Readiness Indicator file check\\\\n2025-10-07T22:09:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.125551 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.144091 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.164230 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.181282 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.181369 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.181397 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.181433 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.181459 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:17Z","lastTransitionTime":"2025-10-07T22:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.195976 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:10:08Z\\\",\\\"message\\\":\\\"rom k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.096778 6974 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.096997 6974 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.097343 6974 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.097617 6974 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.098030 6974 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:10:08.098085 6974 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 22:10:08.098102 6974 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:10:08.098107 6974 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:10:08.098157 6974 factory.go:656] Stopping watch factory\\\\nI1007 22:10:08.098178 6974 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:10:08.098209 6974 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:10:08.098217 6974 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 22:10:08.098224 6974 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:10:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.215279 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.228981 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.251164 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.270086 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.285024 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.285087 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.285100 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.285120 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.285136 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:17Z","lastTransitionTime":"2025-10-07T22:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.294918 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.311846 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcfed2fc-23c7-4293-9d0f-b1fe9ac20394\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62ddf0800fa21c83bb765ca4ae13168e4ba543ae0b9ae2c52a264b019b17561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.330836 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.352337 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:17Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.389102 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.389172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.389197 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.389227 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.389248 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:17Z","lastTransitionTime":"2025-10-07T22:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.494428 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.494569 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.494590 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.494654 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.494672 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:17Z","lastTransitionTime":"2025-10-07T22:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.598523 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.598579 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.598594 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.598616 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.598630 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:17Z","lastTransitionTime":"2025-10-07T22:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.702171 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.702633 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.702777 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.702950 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.703091 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:17Z","lastTransitionTime":"2025-10-07T22:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.806333 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.806420 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.806447 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.806485 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.806507 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:17Z","lastTransitionTime":"2025-10-07T22:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.909199 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.909278 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.909300 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.909329 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.909348 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:17Z","lastTransitionTime":"2025-10-07T22:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.982242 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.982349 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:17 crc kubenswrapper[4871]: E1007 22:10:17.982511 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:17 crc kubenswrapper[4871]: I1007 22:10:17.982527 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:17 crc kubenswrapper[4871]: E1007 22:10:17.982859 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:17 crc kubenswrapper[4871]: E1007 22:10:17.982955 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.012937 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.012983 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.013001 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.013028 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.013046 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.116738 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.116841 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.116861 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.116891 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.116912 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.219862 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.219943 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.219961 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.219987 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.220003 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.323182 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.323237 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.323247 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.323267 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.323278 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.427358 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.427417 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.427430 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.427448 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.427463 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.531141 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.531202 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.531219 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.531244 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.531261 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.634173 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.634231 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.634257 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.634288 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.634309 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.737190 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.737239 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.737257 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.737277 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.737292 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.839639 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.839700 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.839723 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.839752 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.839774 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.942204 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.942271 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.942290 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.942312 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.942330 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:18Z","lastTransitionTime":"2025-10-07T22:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:18 crc kubenswrapper[4871]: I1007 22:10:18.982142 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:18 crc kubenswrapper[4871]: E1007 22:10:18.982404 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.045399 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.045526 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.045550 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.045576 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.045596 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.148932 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.149009 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.149034 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.149068 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.149092 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.251628 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.251695 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.251780 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.251866 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.251889 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.355845 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.355907 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.355921 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.356002 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.356019 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.459186 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.459257 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.459276 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.459304 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.459323 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.562916 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.562986 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.563011 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.563043 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.563069 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.665976 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.666062 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.666088 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.666127 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.666153 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.768696 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.768746 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.768756 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.768773 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.768785 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.871522 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.871585 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.871821 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.871858 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.871888 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.974780 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.974879 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.974899 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.974931 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.974956 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:19Z","lastTransitionTime":"2025-10-07T22:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.982128 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.982185 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:19 crc kubenswrapper[4871]: I1007 22:10:19.982362 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:19 crc kubenswrapper[4871]: E1007 22:10:19.982575 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:19 crc kubenswrapper[4871]: E1007 22:10:19.983143 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:19 crc kubenswrapper[4871]: E1007 22:10:19.983328 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.078070 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.078115 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.078131 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.078155 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.078173 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:20Z","lastTransitionTime":"2025-10-07T22:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.181052 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.181154 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.181181 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.181217 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.181245 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:20Z","lastTransitionTime":"2025-10-07T22:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.284107 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.284167 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.284187 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.284216 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.284236 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:20Z","lastTransitionTime":"2025-10-07T22:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.387339 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.387389 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.387411 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.387448 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.387464 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:20Z","lastTransitionTime":"2025-10-07T22:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.491215 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.491273 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.491296 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.491323 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.491344 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:20Z","lastTransitionTime":"2025-10-07T22:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.595086 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.595162 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.595183 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.595210 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.595228 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:20Z","lastTransitionTime":"2025-10-07T22:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.699039 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.699110 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.699132 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.699160 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.699180 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:20Z","lastTransitionTime":"2025-10-07T22:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.803229 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.803302 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.803327 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.803357 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.803379 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:20Z","lastTransitionTime":"2025-10-07T22:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.906612 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.906691 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.906715 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.906751 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.906775 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:20Z","lastTransitionTime":"2025-10-07T22:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:20 crc kubenswrapper[4871]: I1007 22:10:20.981670 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:20 crc kubenswrapper[4871]: E1007 22:10:20.982011 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.010351 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.010418 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.010442 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.010466 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.010487 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.114450 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.114523 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.114547 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.114582 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.114610 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.218661 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.218740 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.218758 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.218819 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.218845 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.322446 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.322524 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.322547 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.322579 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.322599 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.426362 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.426419 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.426437 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.426465 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.426484 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.530541 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.530596 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.530613 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.530637 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.530654 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.634068 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.634147 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.634164 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.634194 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.634213 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.737878 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.737957 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.737981 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.738012 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.738035 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.845625 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.845698 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.845720 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.845752 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.845776 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.949085 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.949142 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.949161 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.949188 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.949205 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:21Z","lastTransitionTime":"2025-10-07T22:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.981440 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.981566 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.981571 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:21 crc kubenswrapper[4871]: E1007 22:10:21.981691 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:21 crc kubenswrapper[4871]: E1007 22:10:21.981744 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:21 crc kubenswrapper[4871]: E1007 22:10:21.982340 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:21 crc kubenswrapper[4871]: I1007 22:10:21.982709 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:10:21 crc kubenswrapper[4871]: E1007 22:10:21.982994 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.052745 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.052873 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.052897 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.052932 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.053011 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.157907 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.157968 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.157986 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.158008 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.158027 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.267849 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.267911 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.267931 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.267959 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.267979 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.370890 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.370965 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.370982 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.371010 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.371028 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.478964 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.479110 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.479136 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.479168 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.479193 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.582430 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.582488 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.582508 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.582535 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.582551 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.685084 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.685149 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.685165 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.685195 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.685215 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.788382 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.788457 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.788473 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.788497 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.788512 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.892193 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.892282 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.892310 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.892345 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.892370 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.981520 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:22 crc kubenswrapper[4871]: E1007 22:10:22.981971 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.994993 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.995062 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.995087 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.995121 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:22 crc kubenswrapper[4871]: I1007 22:10:22.995145 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:22Z","lastTransitionTime":"2025-10-07T22:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.098249 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.098330 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.098358 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.098386 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.098408 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:23Z","lastTransitionTime":"2025-10-07T22:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.202085 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.202154 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.202180 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.202208 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.202226 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:23Z","lastTransitionTime":"2025-10-07T22:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.305577 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.305627 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.305638 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.305656 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.305666 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:23Z","lastTransitionTime":"2025-10-07T22:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.408656 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.408739 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.408760 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.408818 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.408841 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:23Z","lastTransitionTime":"2025-10-07T22:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.512439 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.512510 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.512529 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.512554 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.512573 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:23Z","lastTransitionTime":"2025-10-07T22:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.616487 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.616578 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.616602 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.616637 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.616662 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:23Z","lastTransitionTime":"2025-10-07T22:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.720078 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.720140 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.720157 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.720184 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.720205 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:23Z","lastTransitionTime":"2025-10-07T22:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.823584 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.823711 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.823731 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.823758 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.823777 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:23Z","lastTransitionTime":"2025-10-07T22:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.930458 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.930522 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.930538 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.930565 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.930582 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:23Z","lastTransitionTime":"2025-10-07T22:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.981298 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.981375 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:23 crc kubenswrapper[4871]: E1007 22:10:23.981483 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:23 crc kubenswrapper[4871]: I1007 22:10:23.981298 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:23 crc kubenswrapper[4871]: E1007 22:10:23.981680 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:23 crc kubenswrapper[4871]: E1007 22:10:23.982045 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.034636 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.034716 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.034735 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.034763 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.034780 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.137731 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.137856 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.137875 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.137901 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.137920 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.241835 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.241919 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.241938 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.241965 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.241984 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.345196 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.345291 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.345320 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.345355 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.345381 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.449012 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.449075 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.449097 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.449124 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.449144 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.553093 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.553155 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.553172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.553192 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.553206 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.656245 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.656325 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.656339 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.656365 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.656378 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.759596 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.759672 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.759695 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.759725 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.759749 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.863299 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.863371 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.863389 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.863416 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.863436 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.967705 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.967825 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.967853 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.967892 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.967916 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:24Z","lastTransitionTime":"2025-10-07T22:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:24 crc kubenswrapper[4871]: I1007 22:10:24.981946 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:24 crc kubenswrapper[4871]: E1007 22:10:24.982276 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.071669 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.071752 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.071780 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.071848 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.071871 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.174876 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.174962 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.174984 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.175021 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.175045 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.278213 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.278249 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.278267 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.278286 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.278299 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.382306 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.382351 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.382362 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.382381 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.382395 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.485082 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.485168 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.485192 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.485223 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.485245 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.597004 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.597065 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.597082 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.597105 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.597120 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.701143 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.701198 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.701217 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.701245 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.701265 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.804902 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.804976 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.804996 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.805025 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.805043 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.909576 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.909630 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.909651 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.909676 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.909697 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.982212 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.982310 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.982212 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:25 crc kubenswrapper[4871]: E1007 22:10:25.982463 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:25 crc kubenswrapper[4871]: E1007 22:10:25.982570 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:25 crc kubenswrapper[4871]: E1007 22:10:25.982851 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.987107 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.987161 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.987181 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.987206 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:25 crc kubenswrapper[4871]: I1007 22:10:25.987224 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:25Z","lastTransitionTime":"2025-10-07T22:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: E1007 22:10:26.008774 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:26Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.014250 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.014317 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.014335 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.014360 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.014378 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: E1007 22:10:26.034664 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:26Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.040240 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.040294 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.040313 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.040340 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.040359 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: E1007 22:10:26.059972 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:26Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.065203 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.065271 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.065293 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.065331 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.065353 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: E1007 22:10:26.087678 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:26Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.093579 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.093632 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.093650 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.093674 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.093694 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: E1007 22:10:26.112584 4871 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T22:10:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78aa46f6-3267-4ce5-ade1-72b92c4bf3c8\\\",\\\"systemUUID\\\":\\\"32db1e84-0947-42a3-be43-1cbd898711cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:26Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:26 crc kubenswrapper[4871]: E1007 22:10:26.112840 4871 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.114886 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.114935 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.114955 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.114978 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.114996 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.219066 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.219131 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.219148 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.219174 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.219193 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.322826 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.322909 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.322928 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.322955 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.322974 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.426278 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.426353 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.426372 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.426398 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.426416 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.530401 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.530485 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.530505 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.530536 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.530559 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.634259 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.634326 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.634344 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.634372 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.634398 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.737772 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.737880 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.737900 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.737939 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.737961 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.841300 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.841376 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.841397 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.841431 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.841455 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.945679 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.945743 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.945767 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.945831 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.945857 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:26Z","lastTransitionTime":"2025-10-07T22:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.946873 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:26 crc kubenswrapper[4871]: E1007 22:10:26.947157 4871 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:10:26 crc kubenswrapper[4871]: E1007 22:10:26.947301 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs podName:684fd639-9051-4c1a-8bf0-9d1cb7b5cf59 nodeName:}" failed. No retries permitted until 2025-10-07 22:11:30.947259326 +0000 UTC m=+164.749957589 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs") pod "network-metrics-daemon-gbxdg" (UID: "684fd639-9051-4c1a-8bf0-9d1cb7b5cf59") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 22:10:26 crc kubenswrapper[4871]: I1007 22:10:26.981896 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:26 crc kubenswrapper[4871]: E1007 22:10:26.982360 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.007510 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7add043f-9ee6-4543-9678-a9054c1dadbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a72065bd5bbe4d0722c54e5d68cd567968af7a1956695b14ee36be75dc139e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d890cb6735d341ed1a2c7b29e1bfc646ea7e8956c07e2bf22bf25af67baaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37cc565dc56564ea773ee291ae57e29737db72557bc9dfac3096af613f45947\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9b1660c857a157c6887ba8b5ddd731f1beded6a731aa9369585a609d6b1cd38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86b0010369b9f3cce288c8a5ab29133d5c156d467480453deadcbe878e3a4d23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW1007 22:09:07.100931 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1007 22:09:07.101190 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 22:09:07.101990 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1020120214/tls.crt::/tmp/serving-cert-1020120214/tls.key\\\\\\\"\\\\nI1007 22:09:07.388222 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1007 22:09:07.390991 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1007 22:09:07.391011 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1007 22:09:07.391041 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1007 22:09:07.391046 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1007 22:09:07.396117 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1007 22:09:07.396130 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1007 22:09:07.396153 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396163 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1007 22:09:07.396173 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1007 22:09:07.396177 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1007 22:09:07.396182 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1007 22:09:07.396185 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1007 22:09:07.399252 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ee43ef1cd6aa77725719f1764fc842e38e53cb5b32d6c8ce08fe7c9c1af459\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03733e82463a9e48e8babb9a031b16be330b768b5e3bdd982fbf453c5ae3f0be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.030453 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1db71aa65d644e55c9ba22921b1baf26855679c957ba556d1dd64980139517e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1482bebc5c04ddbf6924efce3ef0ee5edca3099b2567bfa2316be97aa3be46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.049454 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae3279b09b76a137e883919c2951804f3c8cf3a5b5658253f6721590c9cdd44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.050097 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.050153 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.050172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.050200 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.050221 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.069753 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5008be0-9415-428f-917e-7ec0905be00d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b2790246f39181a28261f18206f639b83ffb71a74e3dff89407e575d4c3b7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bdb905a8b41ca011ec0414a4aaf5b9545cbc6389788c0f168288aa87a585cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xkz4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:21Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hlhw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.089199 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcfed2fc-23c7-4293-9d0f-b1fe9ac20394\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62ddf0800fa21c83bb765ca4ae13168e4ba543ae0b9ae2c52a264b019b17561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641e9d364fd4fcb32f9d79e0134ad9acc82ed340e3f02781168ce6026d59b30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.112684 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff2ba72d7be262a8149516e46097bee105fc6b621dc942cebc9c55fa229637f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.137191 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.155969 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.156055 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.156078 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.156110 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.156131 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.160326 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef3b09cf-b090-4f2c-892d-ab7f7aee3129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4237b2a9f57147f449838ad0238a3b9d0d1984c97cedc4cd311355e76e31c14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8x84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hp4jl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.180238 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xsct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gbxdg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.200639 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43152056-ae0e-4907-987c-7b907a40ce52\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36bdd168808a10c793ae294af80713b245cda9c3ff33bbb98f124bca9bb71e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e54b9a3054c7d53cbaf5f2fc6dc05f341f3666ff0c5f55a5634b2765665bb1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://683b8182e06506f39085daabaefbb83f5d8f676400cb0e4798e4636019a4be2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adb04670457eade87a4ae65ebdc50d6c923b57b57dad9b65e908597a4e1d44c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.220983 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.240539 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.260260 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.260419 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.260448 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.260527 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.260552 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.264871 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nv4jc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53708429-59b4-4319-bdb7-8a922a551e59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:09:55Z\\\",\\\"message\\\":\\\"2025-10-07T22:09:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56\\\\n2025-10-07T22:09:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a62df7bd-6c89-497b-9c13-e70697ea0f56 to /host/opt/cni/bin/\\\\n2025-10-07T22:09:10Z [verbose] multus-daemon started\\\\n2025-10-07T22:09:10Z [verbose] Readiness Indicator file check\\\\n2025-10-07T22:09:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2q9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nv4jc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.284165 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xwhnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d535a1fa-1557-4e32-8583-d8415c47c1bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ed7c59f44aacc6575313fd51b2536352567f9319a519ea3137ef67e497bb9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jk8nx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:11Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xwhnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.303858 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl6pr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"817bca9c-66e5-440f-a773-d49006702b3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://939e9a16a0bbe7729e8172675d5ba8ce353b4b522698f88a046ab072eaa177e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsnsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl6pr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.329022 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h66dd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad179ea2-56e8-4e5e-a72c-599c797918d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a8b1686d9c39ab23cf0de1a79ed75f290797cb5cafe1328b3c0a6e20dd9dbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31dbdc90c2ebf28ad2457848c6c22eaae3e33c667b55282d43c63f4f5f4bf77f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da2308e041bd27a021c002f2cb92b3e3833694130566ac6706e7eb194a4f2bde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://018b0ebe8da45ddad35388cdbeca09d08f443bad7f254225544664e87b2253ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95f984c15372eb9a7e1beaff049012d977a120760595cb8ba4fac5a2a21bb8b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://029f9a94348419e8f763934b35071d46e5b08823c530f6a9124f2e5d78cbbd13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6f153545203ad6d5e9295f793c3c89d8b83176e2389c983fa9572ffe5c52bb5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shmsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h66dd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.363884 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.363942 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.363961 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.363986 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.364006 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.365202 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0059e7e8-7d63-4b03-81a0-b8521803f34a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T22:10:08Z\\\",\\\"message\\\":\\\"rom k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.096778 6974 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.096997 6974 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.097343 6974 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.097617 6974 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 22:10:08.098030 6974 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 22:10:08.098085 6974 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 22:10:08.098102 6974 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 22:10:08.098107 6974 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 22:10:08.098157 6974 factory.go:656] Stopping watch factory\\\\nI1007 22:10:08.098178 6974 ovnkube.go:599] Stopped ovnkube\\\\nI1007 22:10:08.098209 6974 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 22:10:08.098217 6974 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 22:10:08.098224 6974 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 22\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T22:10:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T22:09:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T22:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7hp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:09:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-d44j5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.388515 4871 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a2f713e-3904-4672-9623-e2c8d266967c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:09:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T22:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10898693022cd9510bfe7231598380f73dd5546cc287fc4c201df126af7f46a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e523af8811f0113e52cc33b312086f6a7ca9a7580b2390eaff6006a02006593a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825e98878d4292b9ab255af1fd0b2a076e0e33d035a507c0cc7a394413626233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3760723e385a9fc149ae2eece64daf8d58271d16dcd2b23a07fe87dc086f540f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T22:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T22:08:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T22:10:27Z is after 2025-08-24T17:21:41Z" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.467485 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.467551 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.467569 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.467597 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.467615 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.571166 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.571518 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.571654 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.571850 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.571975 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.676077 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.676145 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.676163 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.676192 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.676210 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.779232 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.779297 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.779316 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.779343 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.779363 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.882031 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.882116 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.882134 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.882165 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.882185 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.981609 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.981754 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:27 crc kubenswrapper[4871]: E1007 22:10:27.981833 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.981607 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:27 crc kubenswrapper[4871]: E1007 22:10:27.982018 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:27 crc kubenswrapper[4871]: E1007 22:10:27.982187 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.984864 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.984942 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.984963 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.984987 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:27 crc kubenswrapper[4871]: I1007 22:10:27.985009 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:27Z","lastTransitionTime":"2025-10-07T22:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.088342 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.088412 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.088432 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.088461 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.088479 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:28Z","lastTransitionTime":"2025-10-07T22:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.192090 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.192162 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.192186 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.192333 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.192374 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:28Z","lastTransitionTime":"2025-10-07T22:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.296068 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.296172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.296196 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.296224 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.296243 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:28Z","lastTransitionTime":"2025-10-07T22:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.400266 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.400346 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.400554 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.400587 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.400609 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:28Z","lastTransitionTime":"2025-10-07T22:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.503827 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.503893 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.503911 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.503939 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.503958 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:28Z","lastTransitionTime":"2025-10-07T22:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.607721 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.607828 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.607848 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.607877 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.607899 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:28Z","lastTransitionTime":"2025-10-07T22:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.711509 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.711569 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.711592 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.711619 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.711637 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:28Z","lastTransitionTime":"2025-10-07T22:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.814892 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.814976 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.814994 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.815024 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.815044 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:28Z","lastTransitionTime":"2025-10-07T22:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.918703 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.918753 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.918767 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.918785 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.918817 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:28Z","lastTransitionTime":"2025-10-07T22:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:28 crc kubenswrapper[4871]: I1007 22:10:28.981878 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:28 crc kubenswrapper[4871]: E1007 22:10:28.982066 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.021313 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.021395 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.021414 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.021443 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.021466 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.124593 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.124633 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.124648 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.124670 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.124682 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.228292 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.228341 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.228352 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.228370 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.228383 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.331695 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.331765 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.331831 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.331864 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.331885 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.435608 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.435642 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.435652 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.435679 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.435691 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.538925 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.538997 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.539020 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.539057 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.539081 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.643286 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.643326 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.643337 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.643355 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.643367 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.747053 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.747133 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.747153 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.747183 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.747208 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.850716 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.850834 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.850860 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.850893 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.850918 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.954163 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.954211 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.954223 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.954240 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.954251 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:29Z","lastTransitionTime":"2025-10-07T22:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.982192 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.982257 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:29 crc kubenswrapper[4871]: I1007 22:10:29.982282 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:29 crc kubenswrapper[4871]: E1007 22:10:29.982389 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:29 crc kubenswrapper[4871]: E1007 22:10:29.982526 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:29 crc kubenswrapper[4871]: E1007 22:10:29.982640 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.058524 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.058597 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.058615 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.058646 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.058667 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.162231 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.162623 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.162841 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.163349 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.163546 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.267715 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.268944 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.269126 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.269280 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.269442 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.373559 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.373644 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.373664 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.373693 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.373715 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.477952 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.478038 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.478061 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.478096 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.478119 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.582007 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.582055 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.582066 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.582086 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.582098 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.686319 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.686378 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.686406 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.686437 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.686462 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.789855 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.789930 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.789953 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.789984 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.790007 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.892918 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.892980 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.892997 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.893024 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.893042 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.982421 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:30 crc kubenswrapper[4871]: E1007 22:10:30.982764 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.996607 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.996682 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.996716 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.996752 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:30 crc kubenswrapper[4871]: I1007 22:10:30.996774 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:30Z","lastTransitionTime":"2025-10-07T22:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.005961 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.100314 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.100418 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.100437 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.100464 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.100489 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:31Z","lastTransitionTime":"2025-10-07T22:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.203769 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.203865 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.203891 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.203920 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.203938 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:31Z","lastTransitionTime":"2025-10-07T22:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.307457 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.307534 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.307551 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.307571 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.307585 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:31Z","lastTransitionTime":"2025-10-07T22:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.411210 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.411259 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.411271 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.411292 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.411307 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:31Z","lastTransitionTime":"2025-10-07T22:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.515358 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.515406 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.515418 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.515435 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.515447 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:31Z","lastTransitionTime":"2025-10-07T22:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.619302 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.619395 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.619413 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.619438 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.619456 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:31Z","lastTransitionTime":"2025-10-07T22:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.723096 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.723172 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.723191 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.723220 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.723240 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:31Z","lastTransitionTime":"2025-10-07T22:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.827181 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.827254 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.827273 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.827313 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.827333 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:31Z","lastTransitionTime":"2025-10-07T22:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.932319 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.932377 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.932394 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.932416 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.932430 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:31Z","lastTransitionTime":"2025-10-07T22:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.981330 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:31 crc kubenswrapper[4871]: E1007 22:10:31.981559 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.981716 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:31 crc kubenswrapper[4871]: I1007 22:10:31.981730 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:31 crc kubenswrapper[4871]: E1007 22:10:31.982224 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:31 crc kubenswrapper[4871]: E1007 22:10:31.982268 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.036607 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.036672 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.036693 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.036719 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.036738 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.140067 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.140142 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.140161 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.140196 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.140217 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.244560 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.244650 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.244668 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.244696 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.244722 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.348746 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.348842 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.348860 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.348887 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.348905 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.452770 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.452844 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.452867 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.452896 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.452918 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.556238 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.556287 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.556303 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.556328 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.556345 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.659311 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.659393 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.659411 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.659441 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.659461 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.762625 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.762703 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.762723 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.762753 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.762774 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.866513 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.866574 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.866595 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.866623 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.866643 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.970734 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.970834 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.970860 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.970889 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.970909 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:32Z","lastTransitionTime":"2025-10-07T22:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:32 crc kubenswrapper[4871]: I1007 22:10:32.982239 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:32 crc kubenswrapper[4871]: E1007 22:10:32.982424 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.074856 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.074930 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.074953 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.074986 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.075010 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:33Z","lastTransitionTime":"2025-10-07T22:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.178811 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.178849 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.178863 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.178883 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.178896 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:33Z","lastTransitionTime":"2025-10-07T22:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.281995 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.282078 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.282099 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.282127 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.282151 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:33Z","lastTransitionTime":"2025-10-07T22:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.385628 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.385697 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.385716 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.385749 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.385769 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:33Z","lastTransitionTime":"2025-10-07T22:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.489036 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.489089 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.489102 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.489122 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.489135 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:33Z","lastTransitionTime":"2025-10-07T22:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.592129 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.592195 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.592215 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.592242 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.592260 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:33Z","lastTransitionTime":"2025-10-07T22:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.695553 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.695616 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.695642 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.695673 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.695696 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:33Z","lastTransitionTime":"2025-10-07T22:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.799248 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.799314 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.799342 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.799369 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.799395 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:33Z","lastTransitionTime":"2025-10-07T22:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.902274 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.902345 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.902366 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.902394 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.902413 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:33Z","lastTransitionTime":"2025-10-07T22:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.981740 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.981880 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:33 crc kubenswrapper[4871]: I1007 22:10:33.981740 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:33 crc kubenswrapper[4871]: E1007 22:10:33.982002 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:33 crc kubenswrapper[4871]: E1007 22:10:33.982202 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:33 crc kubenswrapper[4871]: E1007 22:10:33.982320 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.006274 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.006366 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.006387 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.006420 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.006444 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.109705 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.109774 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.109866 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.109904 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.109927 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.235472 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.235549 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.235567 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.235590 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.235607 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.339009 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.339081 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.339102 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.339129 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.339151 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.442369 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.442444 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.442469 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.442503 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.442530 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.545977 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.546060 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.546080 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.546107 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.546127 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.649302 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.649387 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.649400 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.649421 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.649432 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.760779 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.760882 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.760906 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.760938 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.760963 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.864056 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.864132 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.864151 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.864183 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.864203 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.967162 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.967217 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.967235 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.967260 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.967279 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:34Z","lastTransitionTime":"2025-10-07T22:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.981546 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:34 crc kubenswrapper[4871]: E1007 22:10:34.981976 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:34 crc kubenswrapper[4871]: I1007 22:10:34.982228 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:10:34 crc kubenswrapper[4871]: E1007 22:10:34.982433 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.070957 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.071006 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.071018 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.071037 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.071052 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:35Z","lastTransitionTime":"2025-10-07T22:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.173668 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.173717 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.173730 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.173750 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.173764 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:35Z","lastTransitionTime":"2025-10-07T22:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.277603 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.277666 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.277686 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.277720 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.277747 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:35Z","lastTransitionTime":"2025-10-07T22:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.382033 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.382150 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.382235 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.382283 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.382344 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:35Z","lastTransitionTime":"2025-10-07T22:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.485349 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.485423 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.485443 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.485472 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.485491 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:35Z","lastTransitionTime":"2025-10-07T22:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.588376 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.588424 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.588440 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.588466 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.588483 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:35Z","lastTransitionTime":"2025-10-07T22:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.691278 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.691334 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.691349 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.691375 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.691393 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:35Z","lastTransitionTime":"2025-10-07T22:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.794921 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.795003 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.795025 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.795054 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.795076 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:35Z","lastTransitionTime":"2025-10-07T22:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.899031 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.899115 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.899139 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.899176 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.899203 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:35Z","lastTransitionTime":"2025-10-07T22:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.982387 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.982433 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:35 crc kubenswrapper[4871]: E1007 22:10:35.982621 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:35 crc kubenswrapper[4871]: I1007 22:10:35.982762 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:35 crc kubenswrapper[4871]: E1007 22:10:35.982981 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:35 crc kubenswrapper[4871]: E1007 22:10:35.983107 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.002446 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.002536 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.002554 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.002580 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.002599 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:36Z","lastTransitionTime":"2025-10-07T22:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.105657 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.105718 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.105737 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.105761 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.105779 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:36Z","lastTransitionTime":"2025-10-07T22:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.210132 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.210226 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.210257 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.210293 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.210320 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:36Z","lastTransitionTime":"2025-10-07T22:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.313220 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.313292 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.313312 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.313341 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.313363 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:36Z","lastTransitionTime":"2025-10-07T22:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.417505 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.417580 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.417601 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.417627 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.417646 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:36Z","lastTransitionTime":"2025-10-07T22:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.512354 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.512442 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.512461 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.512488 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.512507 4871 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T22:10:36Z","lastTransitionTime":"2025-10-07T22:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.589401 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw"] Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.590055 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.593090 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.593308 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.596357 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.598519 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.623516 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-h66dd" podStartSLOduration=88.623431381 podStartE2EDuration="1m28.623431381s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:36.623131013 +0000 UTC m=+110.425829106" watchObservedRunningTime="2025-10-07 22:10:36.623431381 +0000 UTC m=+110.426129494" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.664457 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.664564 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.664722 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.664760 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.664858 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.717656 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=89.71762649 podStartE2EDuration="1m29.71762649s" podCreationTimestamp="2025-10-07 22:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:36.699191569 +0000 UTC m=+110.501889672" watchObservedRunningTime="2025-10-07 22:10:36.71762649 +0000 UTC m=+110.520324603" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.717889 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-vl6pr" podStartSLOduration=88.717881697 podStartE2EDuration="1m28.717881697s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:36.717593469 +0000 UTC m=+110.520291552" watchObservedRunningTime="2025-10-07 22:10:36.717881697 +0000 UTC m=+110.520579810" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.765824 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.765917 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.765953 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.766021 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.766056 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.766155 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.766874 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.767672 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.779284 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.779469 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hlhw8" podStartSLOduration=87.779448317 podStartE2EDuration="1m27.779448317s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:36.779170209 +0000 UTC m=+110.581868332" watchObservedRunningTime="2025-10-07 22:10:36.779448317 +0000 UTC m=+110.582146420" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.789962 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-t9fcw\" (UID: \"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.821737 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=40.821694032 podStartE2EDuration="40.821694032s" podCreationTimestamp="2025-10-07 22:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:36.799262554 +0000 UTC m=+110.601960697" watchObservedRunningTime="2025-10-07 22:10:36.821694032 +0000 UTC m=+110.624392145" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.822248 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.822234016 podStartE2EDuration="1m28.822234016s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:36.821603649 +0000 UTC m=+110.624301772" watchObservedRunningTime="2025-10-07 22:10:36.822234016 +0000 UTC m=+110.624932129" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.887420 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podStartSLOduration=88.887387942 podStartE2EDuration="1m28.887387942s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:36.886988101 +0000 UTC m=+110.689686184" watchObservedRunningTime="2025-10-07 22:10:36.887387942 +0000 UTC m=+110.690086025" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.912544 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.958293 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=54.95826979 podStartE2EDuration="54.95826979s" podCreationTimestamp="2025-10-07 22:09:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:36.936679605 +0000 UTC m=+110.739377678" watchObservedRunningTime="2025-10-07 22:10:36.95826979 +0000 UTC m=+110.760967863" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.981315 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:36 crc kubenswrapper[4871]: E1007 22:10:36.981434 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:36 crc kubenswrapper[4871]: I1007 22:10:36.992008 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-nv4jc" podStartSLOduration=88.991976638 podStartE2EDuration="1m28.991976638s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:36.991361311 +0000 UTC m=+110.794059384" watchObservedRunningTime="2025-10-07 22:10:36.991976638 +0000 UTC m=+110.794674711" Oct 07 22:10:37 crc kubenswrapper[4871]: I1007 22:10:37.004956 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-xwhnv" podStartSLOduration=89.004931223 podStartE2EDuration="1m29.004931223s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:37.004144792 +0000 UTC m=+110.806842865" watchObservedRunningTime="2025-10-07 22:10:37.004931223 +0000 UTC m=+110.807629296" Oct 07 22:10:37 crc kubenswrapper[4871]: I1007 22:10:37.049892 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=7.049867109 podStartE2EDuration="7.049867109s" podCreationTimestamp="2025-10-07 22:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:37.044176938 +0000 UTC m=+110.846875011" watchObservedRunningTime="2025-10-07 22:10:37.049867109 +0000 UTC m=+110.852565202" Oct 07 22:10:37 crc kubenswrapper[4871]: I1007 22:10:37.723349 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" event={"ID":"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4","Type":"ContainerStarted","Data":"d36c296362000c15f9453bb9d3af5bc08ad95ef6bf54b75027eae272c9bf4576"} Oct 07 22:10:37 crc kubenswrapper[4871]: I1007 22:10:37.723443 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" event={"ID":"1d6a7e4d-b110-4e3f-8ef9-dd0cf7b9a2a4","Type":"ContainerStarted","Data":"1aa1d1e1d67889b1bd5d34de6126e2f27b9b66c5404c072ad3214aaa27d94f1a"} Oct 07 22:10:37 crc kubenswrapper[4871]: I1007 22:10:37.981436 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:37 crc kubenswrapper[4871]: I1007 22:10:37.981504 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:37 crc kubenswrapper[4871]: I1007 22:10:37.981536 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:37 crc kubenswrapper[4871]: E1007 22:10:37.981618 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:37 crc kubenswrapper[4871]: E1007 22:10:37.981729 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:37 crc kubenswrapper[4871]: E1007 22:10:37.981989 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:38 crc kubenswrapper[4871]: I1007 22:10:38.981743 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:38 crc kubenswrapper[4871]: E1007 22:10:38.982043 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:39 crc kubenswrapper[4871]: I1007 22:10:39.981742 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:39 crc kubenswrapper[4871]: I1007 22:10:39.981862 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:39 crc kubenswrapper[4871]: I1007 22:10:39.981742 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:39 crc kubenswrapper[4871]: E1007 22:10:39.981972 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:39 crc kubenswrapper[4871]: E1007 22:10:39.982082 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:39 crc kubenswrapper[4871]: E1007 22:10:39.982220 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:40 crc kubenswrapper[4871]: I1007 22:10:40.982284 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:40 crc kubenswrapper[4871]: E1007 22:10:40.982529 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:41 crc kubenswrapper[4871]: I1007 22:10:41.981655 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:41 crc kubenswrapper[4871]: I1007 22:10:41.981720 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:41 crc kubenswrapper[4871]: E1007 22:10:41.982224 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:41 crc kubenswrapper[4871]: I1007 22:10:41.981726 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:41 crc kubenswrapper[4871]: E1007 22:10:41.982466 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:41 crc kubenswrapper[4871]: E1007 22:10:41.982639 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:42 crc kubenswrapper[4871]: I1007 22:10:42.744839 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nv4jc_53708429-59b4-4319-bdb7-8a922a551e59/kube-multus/1.log" Oct 07 22:10:42 crc kubenswrapper[4871]: I1007 22:10:42.745995 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nv4jc_53708429-59b4-4319-bdb7-8a922a551e59/kube-multus/0.log" Oct 07 22:10:42 crc kubenswrapper[4871]: I1007 22:10:42.746052 4871 generic.go:334] "Generic (PLEG): container finished" podID="53708429-59b4-4319-bdb7-8a922a551e59" containerID="757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d" exitCode=1 Oct 07 22:10:42 crc kubenswrapper[4871]: I1007 22:10:42.746093 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nv4jc" event={"ID":"53708429-59b4-4319-bdb7-8a922a551e59","Type":"ContainerDied","Data":"757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d"} Oct 07 22:10:42 crc kubenswrapper[4871]: I1007 22:10:42.746137 4871 scope.go:117] "RemoveContainer" containerID="ffe754daeadd6f40d491f72c9b699751a02efc293379f8f9578e610cdfa77674" Oct 07 22:10:42 crc kubenswrapper[4871]: I1007 22:10:42.747038 4871 scope.go:117] "RemoveContainer" containerID="757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d" Oct 07 22:10:42 crc kubenswrapper[4871]: E1007 22:10:42.747430 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-nv4jc_openshift-multus(53708429-59b4-4319-bdb7-8a922a551e59)\"" pod="openshift-multus/multus-nv4jc" podUID="53708429-59b4-4319-bdb7-8a922a551e59" Oct 07 22:10:42 crc kubenswrapper[4871]: I1007 22:10:42.775348 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-t9fcw" podStartSLOduration=94.775286917 podStartE2EDuration="1m34.775286917s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:10:37.746243138 +0000 UTC m=+111.548941241" watchObservedRunningTime="2025-10-07 22:10:42.775286917 +0000 UTC m=+116.577985010" Oct 07 22:10:42 crc kubenswrapper[4871]: I1007 22:10:42.981991 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:42 crc kubenswrapper[4871]: E1007 22:10:42.982205 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:43 crc kubenswrapper[4871]: I1007 22:10:43.752946 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nv4jc_53708429-59b4-4319-bdb7-8a922a551e59/kube-multus/1.log" Oct 07 22:10:43 crc kubenswrapper[4871]: I1007 22:10:43.981786 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:43 crc kubenswrapper[4871]: I1007 22:10:43.981881 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:43 crc kubenswrapper[4871]: I1007 22:10:43.981950 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:43 crc kubenswrapper[4871]: E1007 22:10:43.982612 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:43 crc kubenswrapper[4871]: E1007 22:10:43.982707 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:43 crc kubenswrapper[4871]: E1007 22:10:43.982953 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:44 crc kubenswrapper[4871]: I1007 22:10:44.982219 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:44 crc kubenswrapper[4871]: E1007 22:10:44.982437 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:46 crc kubenswrapper[4871]: I1007 22:10:46.006869 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:46 crc kubenswrapper[4871]: I1007 22:10:46.006898 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:46 crc kubenswrapper[4871]: E1007 22:10:46.007065 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:46 crc kubenswrapper[4871]: I1007 22:10:46.007176 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:46 crc kubenswrapper[4871]: E1007 22:10:46.007773 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:46 crc kubenswrapper[4871]: I1007 22:10:46.008405 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:10:46 crc kubenswrapper[4871]: E1007 22:10:46.008305 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:46 crc kubenswrapper[4871]: E1007 22:10:46.008707 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-d44j5_openshift-ovn-kubernetes(0059e7e8-7d63-4b03-81a0-b8521803f34a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" Oct 07 22:10:46 crc kubenswrapper[4871]: E1007 22:10:46.936705 4871 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 07 22:10:46 crc kubenswrapper[4871]: I1007 22:10:46.981507 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:46 crc kubenswrapper[4871]: E1007 22:10:46.983454 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:47 crc kubenswrapper[4871]: E1007 22:10:47.090744 4871 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 22:10:47 crc kubenswrapper[4871]: I1007 22:10:47.982356 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:47 crc kubenswrapper[4871]: I1007 22:10:47.982492 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:47 crc kubenswrapper[4871]: E1007 22:10:47.982586 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:47 crc kubenswrapper[4871]: E1007 22:10:47.982747 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:47 crc kubenswrapper[4871]: I1007 22:10:47.982935 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:47 crc kubenswrapper[4871]: E1007 22:10:47.983076 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:48 crc kubenswrapper[4871]: I1007 22:10:48.982256 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:48 crc kubenswrapper[4871]: E1007 22:10:48.982550 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:49 crc kubenswrapper[4871]: I1007 22:10:49.981845 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:49 crc kubenswrapper[4871]: I1007 22:10:49.981892 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:49 crc kubenswrapper[4871]: I1007 22:10:49.981845 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:49 crc kubenswrapper[4871]: E1007 22:10:49.982087 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:49 crc kubenswrapper[4871]: E1007 22:10:49.982384 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:49 crc kubenswrapper[4871]: E1007 22:10:49.982441 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:50 crc kubenswrapper[4871]: I1007 22:10:50.981346 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:50 crc kubenswrapper[4871]: E1007 22:10:50.981526 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:51 crc kubenswrapper[4871]: I1007 22:10:51.981910 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:51 crc kubenswrapper[4871]: I1007 22:10:51.981910 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:51 crc kubenswrapper[4871]: E1007 22:10:51.982167 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:51 crc kubenswrapper[4871]: E1007 22:10:51.982404 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:51 crc kubenswrapper[4871]: I1007 22:10:51.981943 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:51 crc kubenswrapper[4871]: E1007 22:10:51.982567 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:52 crc kubenswrapper[4871]: E1007 22:10:52.092442 4871 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 22:10:52 crc kubenswrapper[4871]: I1007 22:10:52.982275 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:52 crc kubenswrapper[4871]: E1007 22:10:52.982604 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:53 crc kubenswrapper[4871]: I1007 22:10:53.982591 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:53 crc kubenswrapper[4871]: I1007 22:10:53.982632 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:53 crc kubenswrapper[4871]: E1007 22:10:53.982785 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:53 crc kubenswrapper[4871]: E1007 22:10:53.983303 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:53 crc kubenswrapper[4871]: I1007 22:10:53.983959 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:53 crc kubenswrapper[4871]: E1007 22:10:53.984289 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:54 crc kubenswrapper[4871]: I1007 22:10:54.981904 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:54 crc kubenswrapper[4871]: E1007 22:10:54.982151 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:55 crc kubenswrapper[4871]: I1007 22:10:55.981521 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:55 crc kubenswrapper[4871]: I1007 22:10:55.981571 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:55 crc kubenswrapper[4871]: I1007 22:10:55.981577 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:55 crc kubenswrapper[4871]: E1007 22:10:55.981775 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:55 crc kubenswrapper[4871]: E1007 22:10:55.981949 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:55 crc kubenswrapper[4871]: E1007 22:10:55.982264 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:55 crc kubenswrapper[4871]: I1007 22:10:55.982468 4871 scope.go:117] "RemoveContainer" containerID="757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d" Oct 07 22:10:56 crc kubenswrapper[4871]: I1007 22:10:56.811040 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nv4jc_53708429-59b4-4319-bdb7-8a922a551e59/kube-multus/1.log" Oct 07 22:10:56 crc kubenswrapper[4871]: I1007 22:10:56.811451 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nv4jc" event={"ID":"53708429-59b4-4319-bdb7-8a922a551e59","Type":"ContainerStarted","Data":"603cedf2bf5b244f3d25d0cc6a1e215afab476ce358d9dfdef028d22f6c3f899"} Oct 07 22:10:56 crc kubenswrapper[4871]: I1007 22:10:56.981861 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:56 crc kubenswrapper[4871]: E1007 22:10:56.983062 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:57 crc kubenswrapper[4871]: E1007 22:10:57.093581 4871 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 22:10:57 crc kubenswrapper[4871]: I1007 22:10:57.982065 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:57 crc kubenswrapper[4871]: I1007 22:10:57.982065 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:57 crc kubenswrapper[4871]: I1007 22:10:57.982152 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:57 crc kubenswrapper[4871]: E1007 22:10:57.983196 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:57 crc kubenswrapper[4871]: E1007 22:10:57.983431 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:57 crc kubenswrapper[4871]: E1007 22:10:57.983561 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:58 crc kubenswrapper[4871]: I1007 22:10:58.982290 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:10:58 crc kubenswrapper[4871]: E1007 22:10:58.982727 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:10:59 crc kubenswrapper[4871]: I1007 22:10:59.982010 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:10:59 crc kubenswrapper[4871]: I1007 22:10:59.982028 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:10:59 crc kubenswrapper[4871]: I1007 22:10:59.982519 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:10:59 crc kubenswrapper[4871]: E1007 22:10:59.982758 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:10:59 crc kubenswrapper[4871]: E1007 22:10:59.982992 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:10:59 crc kubenswrapper[4871]: E1007 22:10:59.983125 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:10:59 crc kubenswrapper[4871]: I1007 22:10:59.983299 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:11:00 crc kubenswrapper[4871]: I1007 22:11:00.832751 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/3.log" Oct 07 22:11:00 crc kubenswrapper[4871]: I1007 22:11:00.836554 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerStarted","Data":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} Oct 07 22:11:00 crc kubenswrapper[4871]: I1007 22:11:00.837118 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:11:00 crc kubenswrapper[4871]: I1007 22:11:00.882193 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podStartSLOduration=112.882160547 podStartE2EDuration="1m52.882160547s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:00.880221116 +0000 UTC m=+134.682919189" watchObservedRunningTime="2025-10-07 22:11:00.882160547 +0000 UTC m=+134.684858660" Oct 07 22:11:00 crc kubenswrapper[4871]: I1007 22:11:00.957258 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gbxdg"] Oct 07 22:11:00 crc kubenswrapper[4871]: I1007 22:11:00.957469 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:11:00 crc kubenswrapper[4871]: E1007 22:11:00.957710 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:11:01 crc kubenswrapper[4871]: I1007 22:11:01.981346 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:11:01 crc kubenswrapper[4871]: I1007 22:11:01.981388 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:11:01 crc kubenswrapper[4871]: E1007 22:11:01.982054 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:11:01 crc kubenswrapper[4871]: I1007 22:11:01.981420 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:11:01 crc kubenswrapper[4871]: E1007 22:11:01.982899 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:11:01 crc kubenswrapper[4871]: E1007 22:11:01.983041 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:11:02 crc kubenswrapper[4871]: E1007 22:11:02.095566 4871 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 22:11:02 crc kubenswrapper[4871]: I1007 22:11:02.981588 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:11:02 crc kubenswrapper[4871]: E1007 22:11:02.981773 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:11:03 crc kubenswrapper[4871]: I1007 22:11:03.981673 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:11:03 crc kubenswrapper[4871]: I1007 22:11:03.981766 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:11:03 crc kubenswrapper[4871]: I1007 22:11:03.981836 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:11:03 crc kubenswrapper[4871]: E1007 22:11:03.981932 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:11:03 crc kubenswrapper[4871]: E1007 22:11:03.982095 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:11:03 crc kubenswrapper[4871]: E1007 22:11:03.982319 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:11:04 crc kubenswrapper[4871]: I1007 22:11:04.982509 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:11:04 crc kubenswrapper[4871]: E1007 22:11:04.982764 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:11:05 crc kubenswrapper[4871]: I1007 22:11:05.981619 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:11:05 crc kubenswrapper[4871]: I1007 22:11:05.981696 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:11:05 crc kubenswrapper[4871]: I1007 22:11:05.981721 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:11:05 crc kubenswrapper[4871]: E1007 22:11:05.981813 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 22:11:05 crc kubenswrapper[4871]: E1007 22:11:05.982028 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 22:11:05 crc kubenswrapper[4871]: E1007 22:11:05.982226 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 22:11:06 crc kubenswrapper[4871]: I1007 22:11:06.982255 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:11:06 crc kubenswrapper[4871]: E1007 22:11:06.983605 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbxdg" podUID="684fd639-9051-4c1a-8bf0-9d1cb7b5cf59" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.141321 4871 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.205900 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7lxmn"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.206660 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.207391 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.207984 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.210217 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.210775 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.211180 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.211459 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.213579 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2zz24"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.214267 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.216702 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.217749 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.217915 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5x65s"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.219217 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.219536 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.219707 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.219868 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.227310 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.229535 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.230079 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.230193 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.230922 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.234430 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f5tkt"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.234830 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.235423 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236289 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236365 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236414 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236475 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236586 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236621 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236634 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236705 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236728 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236760 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236887 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236895 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236946 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237009 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237035 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237067 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237091 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236952 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237178 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.236899 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237250 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237296 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237192 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237012 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237428 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237441 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237551 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237636 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237828 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.237856 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.238164 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.238249 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.239682 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.242511 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.242733 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.242932 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.243058 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.243307 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-hsm2n"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.243888 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-hsm2n" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.244293 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8vdbw"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.245007 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.246591 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5gxf8"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247034 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247333 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/554a5462-ff56-4c07-b70b-0ff0f696f988-metrics-tls\") pod \"dns-operator-744455d44c-f5tkt\" (UID: \"554a5462-ff56-4c07-b70b-0ff0f696f988\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247392 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a6b7f864-334b-4d77-a8e4-bf62b614fb80-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247425 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0609d7b1-7443-4bcd-935e-ca2a61a359b7-serving-cert\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247454 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8znbq\" (UniqueName: \"kubernetes.io/projected/252cfed8-c315-49d0-911c-425cfb86fabb-kube-api-access-8znbq\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247480 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252cfed8-c315-49d0-911c-425cfb86fabb-config\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247503 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0312d2c-53d5-4916-9956-418acb3ff0b8-etcd-client\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247526 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0312d2c-53d5-4916-9956-418acb3ff0b8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247552 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-client-ca\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247578 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c892w\" (UniqueName: \"kubernetes.io/projected/a6b7f864-334b-4d77-a8e4-bf62b614fb80-kube-api-access-c892w\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247605 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-etcd-serving-ca\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247631 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbrc9\" (UniqueName: \"kubernetes.io/projected/0609d7b1-7443-4bcd-935e-ca2a61a359b7-kube-api-access-gbrc9\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247656 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4p9r\" (UniqueName: \"kubernetes.io/projected/ecf86dbe-b51a-4228-a5ff-959dfe877387-kube-api-access-l4p9r\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247684 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-config\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247708 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f0312d2c-53d5-4916-9956-418acb3ff0b8-audit-dir\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247733 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-client-ca\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247758 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ecf86dbe-b51a-4228-a5ff-959dfe877387-etcd-client\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247783 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f0312d2c-53d5-4916-9956-418acb3ff0b8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247843 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f0312d2c-53d5-4916-9956-418acb3ff0b8-audit-policies\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247868 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0312d2c-53d5-4916-9956-418acb3ff0b8-serving-cert\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247892 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/252cfed8-c315-49d0-911c-425cfb86fabb-etcd-ca\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247927 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ecf86dbe-b51a-4228-a5ff-959dfe877387-serving-cert\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247956 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce8147-0273-421d-9263-3d4eb06bfcfe-serving-cert\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.247978 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-config\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248002 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248027 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-image-import-ca\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248048 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ecf86dbe-b51a-4228-a5ff-959dfe877387-encryption-config\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248071 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/252cfed8-c315-49d0-911c-425cfb86fabb-etcd-client\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248097 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9pnm\" (UniqueName: \"kubernetes.io/projected/7b559a4d-6250-4309-b703-0d70e4652d12-kube-api-access-h9pnm\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248120 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55ee5574-d8f1-48bd-83c3-7aa62e6366f1-serving-cert\") pod \"openshift-config-operator-7777fb866f-9l8jh\" (UID: \"55ee5574-d8f1-48bd-83c3-7aa62e6366f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248142 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/55ee5574-d8f1-48bd-83c3-7aa62e6366f1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9l8jh\" (UID: \"55ee5574-d8f1-48bd-83c3-7aa62e6366f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248168 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/252cfed8-c315-49d0-911c-425cfb86fabb-etcd-service-ca\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248203 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b4d26eda-8425-4a92-aeb1-418ea15e53fa-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j66lt\" (UID: \"b4d26eda-8425-4a92-aeb1-418ea15e53fa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248228 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b559a4d-6250-4309-b703-0d70e4652d12-config\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248255 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7hkb\" (UniqueName: \"kubernetes.io/projected/55ee5574-d8f1-48bd-83c3-7aa62e6366f1-kube-api-access-d7hkb\") pod \"openshift-config-operator-7777fb866f-9l8jh\" (UID: \"55ee5574-d8f1-48bd-83c3-7aa62e6366f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248278 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a6b7f864-334b-4d77-a8e4-bf62b614fb80-images\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248300 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-config\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248337 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7b559a4d-6250-4309-b703-0d70e4652d12-machine-approver-tls\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248360 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248383 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/252cfed8-c315-49d0-911c-425cfb86fabb-serving-cert\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248406 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b559a4d-6250-4309-b703-0d70e4652d12-auth-proxy-config\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248429 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn2nm\" (UniqueName: \"kubernetes.io/projected/554a5462-ff56-4c07-b70b-0ff0f696f988-kube-api-access-mn2nm\") pod \"dns-operator-744455d44c-f5tkt\" (UID: \"554a5462-ff56-4c07-b70b-0ff0f696f988\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248453 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ecf86dbe-b51a-4228-a5ff-959dfe877387-audit-dir\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248486 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f0312d2c-53d5-4916-9956-418acb3ff0b8-encryption-config\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248513 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b7f864-334b-4d77-a8e4-bf62b614fb80-config\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248537 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvrnv\" (UniqueName: \"kubernetes.io/projected/b4d26eda-8425-4a92-aeb1-418ea15e53fa-kube-api-access-cvrnv\") pod \"cluster-samples-operator-665b6dd947-j66lt\" (UID: \"b4d26eda-8425-4a92-aeb1-418ea15e53fa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248561 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p2jk\" (UniqueName: \"kubernetes.io/projected/03ce8147-0273-421d-9263-3d4eb06bfcfe-kube-api-access-9p2jk\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248588 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqqtl\" (UniqueName: \"kubernetes.io/projected/ff939f3b-6f07-4632-930b-0fa62db4d938-kube-api-access-gqqtl\") pod \"downloads-7954f5f757-hsm2n\" (UID: \"ff939f3b-6f07-4632-930b-0fa62db4d938\") " pod="openshift-console/downloads-7954f5f757-hsm2n" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248612 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ecf86dbe-b51a-4228-a5ff-959dfe877387-node-pullsecrets\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248633 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-audit\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.248655 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkrmg\" (UniqueName: \"kubernetes.io/projected/f0312d2c-53d5-4916-9956-418acb3ff0b8-kube-api-access-wkrmg\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.249427 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.249569 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.249673 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.250068 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-tw5sf"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.271117 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.271395 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.272728 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bwsvx"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.273351 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.274900 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.288616 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.288651 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.288972 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289115 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289129 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289169 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289224 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289590 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289658 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289681 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289745 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289772 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289809 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289822 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.289958 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.290013 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.290032 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.290085 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.290208 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.290276 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.290387 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.290433 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.290533 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.291484 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.291630 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.291742 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.291848 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.291970 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.292108 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.292221 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.292330 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.292437 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.293030 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-r8db4"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.293844 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.294077 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.296761 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.298220 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.298879 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-clflh"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.299415 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.304632 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7lxmn"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.304686 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.307154 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.307265 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.309688 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.314616 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.314673 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.315303 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.320575 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.321159 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.321809 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rfkm7"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.322167 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.322737 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8nrs6"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.323241 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.325097 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.326072 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.331720 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.332308 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.332817 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.334096 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.335027 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.335675 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f5tkt"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.335710 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.348646 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.349110 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.349895 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.350457 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.350762 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.351265 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.353952 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f0312d2c-53d5-4916-9956-418acb3ff0b8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.353990 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354013 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f0312d2c-53d5-4916-9956-418acb3ff0b8-audit-policies\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354030 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0312d2c-53d5-4916-9956-418acb3ff0b8-serving-cert\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354050 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/252cfed8-c315-49d0-911c-425cfb86fabb-etcd-ca\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354080 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ecf86dbe-b51a-4228-a5ff-959dfe877387-serving-cert\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354096 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce8147-0273-421d-9263-3d4eb06bfcfe-serving-cert\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354110 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-config\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354124 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354141 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354157 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-image-import-ca\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.354175 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ecf86dbe-b51a-4228-a5ff-959dfe877387-encryption-config\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.355513 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f0312d2c-53d5-4916-9956-418acb3ff0b8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.357473 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-config\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.359307 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.359345 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-image-import-ca\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.359492 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.360111 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/252cfed8-c315-49d0-911c-425cfb86fabb-etcd-ca\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.360231 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f0312d2c-53d5-4916-9956-418acb3ff0b8-audit-policies\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.395505 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.407467 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ecf86dbe-b51a-4228-a5ff-959dfe877387-encryption-config\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408087 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce8147-0273-421d-9263-3d4eb06bfcfe-serving-cert\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408125 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/252cfed8-c315-49d0-911c-425cfb86fabb-etcd-client\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408175 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408218 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9pnm\" (UniqueName: \"kubernetes.io/projected/7b559a4d-6250-4309-b703-0d70e4652d12-kube-api-access-h9pnm\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408246 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-policies\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408282 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55ee5574-d8f1-48bd-83c3-7aa62e6366f1-serving-cert\") pod \"openshift-config-operator-7777fb866f-9l8jh\" (UID: \"55ee5574-d8f1-48bd-83c3-7aa62e6366f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408308 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/55ee5574-d8f1-48bd-83c3-7aa62e6366f1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9l8jh\" (UID: \"55ee5574-d8f1-48bd-83c3-7aa62e6366f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408334 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/252cfed8-c315-49d0-911c-425cfb86fabb-etcd-service-ca\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408363 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b4d26eda-8425-4a92-aeb1-418ea15e53fa-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j66lt\" (UID: \"b4d26eda-8425-4a92-aeb1-418ea15e53fa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408414 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b559a4d-6250-4309-b703-0d70e4652d12-config\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408442 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-config\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408469 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7hkb\" (UniqueName: \"kubernetes.io/projected/55ee5574-d8f1-48bd-83c3-7aa62e6366f1-kube-api-access-d7hkb\") pod \"openshift-config-operator-7777fb866f-9l8jh\" (UID: \"55ee5574-d8f1-48bd-83c3-7aa62e6366f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408500 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a6b7f864-334b-4d77-a8e4-bf62b614fb80-images\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408523 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7b559a4d-6250-4309-b703-0d70e4652d12-machine-approver-tls\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408545 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408583 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/252cfed8-c315-49d0-911c-425cfb86fabb-serving-cert\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408625 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn2nm\" (UniqueName: \"kubernetes.io/projected/554a5462-ff56-4c07-b70b-0ff0f696f988-kube-api-access-mn2nm\") pod \"dns-operator-744455d44c-f5tkt\" (UID: \"554a5462-ff56-4c07-b70b-0ff0f696f988\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408653 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408684 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b559a4d-6250-4309-b703-0d70e4652d12-auth-proxy-config\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408712 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ecf86dbe-b51a-4228-a5ff-959dfe877387-audit-dir\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408734 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f0312d2c-53d5-4916-9956-418acb3ff0b8-encryption-config\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408760 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b7f864-334b-4d77-a8e4-bf62b614fb80-config\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408787 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvrnv\" (UniqueName: \"kubernetes.io/projected/b4d26eda-8425-4a92-aeb1-418ea15e53fa-kube-api-access-cvrnv\") pod \"cluster-samples-operator-665b6dd947-j66lt\" (UID: \"b4d26eda-8425-4a92-aeb1-418ea15e53fa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408851 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqqtl\" (UniqueName: \"kubernetes.io/projected/ff939f3b-6f07-4632-930b-0fa62db4d938-kube-api-access-gqqtl\") pod \"downloads-7954f5f757-hsm2n\" (UID: \"ff939f3b-6f07-4632-930b-0fa62db4d938\") " pod="openshift-console/downloads-7954f5f757-hsm2n" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408876 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p2jk\" (UniqueName: \"kubernetes.io/projected/03ce8147-0273-421d-9263-3d4eb06bfcfe-kube-api-access-9p2jk\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408900 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ecf86dbe-b51a-4228-a5ff-959dfe877387-node-pullsecrets\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408921 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-audit\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408942 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkrmg\" (UniqueName: \"kubernetes.io/projected/f0312d2c-53d5-4916-9956-418acb3ff0b8-kube-api-access-wkrmg\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408967 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/554a5462-ff56-4c07-b70b-0ff0f696f988-metrics-tls\") pod \"dns-operator-744455d44c-f5tkt\" (UID: \"554a5462-ff56-4c07-b70b-0ff0f696f988\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.408991 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a6b7f864-334b-4d77-a8e4-bf62b614fb80-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409012 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-dir\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409038 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0609d7b1-7443-4bcd-935e-ca2a61a359b7-serving-cert\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409062 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8znbq\" (UniqueName: \"kubernetes.io/projected/252cfed8-c315-49d0-911c-425cfb86fabb-kube-api-access-8znbq\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409086 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252cfed8-c315-49d0-911c-425cfb86fabb-config\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409107 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409132 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409157 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0312d2c-53d5-4916-9956-418acb3ff0b8-etcd-client\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409178 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0312d2c-53d5-4916-9956-418acb3ff0b8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409199 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6558\" (UniqueName: \"kubernetes.io/projected/f031a182-f40f-4369-83fa-51ba2f2ebd83-kube-api-access-f6558\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409229 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c892w\" (UniqueName: \"kubernetes.io/projected/a6b7f864-334b-4d77-a8e4-bf62b614fb80-kube-api-access-c892w\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409252 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-client-ca\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409279 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409308 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409333 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-etcd-serving-ca\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409359 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbrc9\" (UniqueName: \"kubernetes.io/projected/0609d7b1-7443-4bcd-935e-ca2a61a359b7-kube-api-access-gbrc9\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409380 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409406 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409432 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4p9r\" (UniqueName: \"kubernetes.io/projected/ecf86dbe-b51a-4228-a5ff-959dfe877387-kube-api-access-l4p9r\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409459 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-config\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409480 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409507 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f0312d2c-53d5-4916-9956-418acb3ff0b8-audit-dir\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409529 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-client-ca\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.409555 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ecf86dbe-b51a-4228-a5ff-959dfe877387-etcd-client\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.410130 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0312d2c-53d5-4916-9956-418acb3ff0b8-serving-cert\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.413127 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.413464 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.423027 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.425041 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ecf86dbe-b51a-4228-a5ff-959dfe877387-serving-cert\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.425557 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/252cfed8-c315-49d0-911c-425cfb86fabb-etcd-client\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.425984 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.427456 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.427487 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.428221 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.428570 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.428784 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.428915 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.429406 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ecf86dbe-b51a-4228-a5ff-959dfe877387-audit-dir\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.430886 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55ee5574-d8f1-48bd-83c3-7aa62e6366f1-serving-cert\") pod \"openshift-config-operator-7777fb866f-9l8jh\" (UID: \"55ee5574-d8f1-48bd-83c3-7aa62e6366f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.432177 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.440536 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/55ee5574-d8f1-48bd-83c3-7aa62e6366f1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9l8jh\" (UID: \"55ee5574-d8f1-48bd-83c3-7aa62e6366f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.446362 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/252cfed8-c315-49d0-911c-425cfb86fabb-etcd-service-ca\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.448641 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-client-ca\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.448823 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.449335 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.449417 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.449592 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.449701 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.449879 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.450065 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.450150 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.450197 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.450280 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.451123 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kvwx2"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.451355 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.451593 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.451652 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.451725 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.452012 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-audit\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.452181 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.452457 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.452486 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.452733 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.452929 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.453005 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.453198 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.453536 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ecf86dbe-b51a-4228-a5ff-959dfe877387-node-pullsecrets\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.453768 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.455208 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.455962 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b559a4d-6250-4309-b703-0d70e4652d12-config\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.456899 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.458982 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f0312d2c-53d5-4916-9956-418acb3ff0b8-encryption-config\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.459898 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.460415 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0609d7b1-7443-4bcd-935e-ca2a61a359b7-serving-cert\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.461173 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-client-ca\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.461459 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a6b7f864-334b-4d77-a8e4-bf62b614fb80-images\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.462163 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b7f864-334b-4d77-a8e4-bf62b614fb80-config\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.462393 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-x5mfq"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.462410 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0312d2c-53d5-4916-9956-418acb3ff0b8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.462688 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252cfed8-c315-49d0-911c-425cfb86fabb-config\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.462830 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.462847 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5gxf8"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.462907 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b559a4d-6250-4309-b703-0d70e4652d12-auth-proxy-config\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.462948 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.463259 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.463722 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-config\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.463886 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ecf86dbe-b51a-4228-a5ff-959dfe877387-etcd-serving-ca\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.463943 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7b559a4d-6250-4309-b703-0d70e4652d12-machine-approver-tls\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.463980 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.463986 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f0312d2c-53d5-4916-9956-418acb3ff0b8-audit-dir\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.464538 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-config\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.464652 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.466015 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/252cfed8-c315-49d0-911c-425cfb86fabb-serving-cert\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.466240 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.466582 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.466744 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.468098 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b4d26eda-8425-4a92-aeb1-418ea15e53fa-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j66lt\" (UID: \"b4d26eda-8425-4a92-aeb1-418ea15e53fa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.469653 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f0312d2c-53d5-4916-9956-418acb3ff0b8-etcd-client\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.469714 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bwsvx"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.469838 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ecf86dbe-b51a-4228-a5ff-959dfe877387-etcd-client\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.471308 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-tw5sf"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.473559 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8z4bc"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.476604 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.476631 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.476752 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.478306 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-clflh"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.479625 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.480300 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a6b7f864-334b-4d77-a8e4-bf62b614fb80-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.481432 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2zz24"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.481818 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/554a5462-ff56-4c07-b70b-0ff0f696f988-metrics-tls\") pod \"dns-operator-744455d44c-f5tkt\" (UID: \"554a5462-ff56-4c07-b70b-0ff0f696f988\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.483146 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.483409 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-hsm2n"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.488039 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.492290 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.499174 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rfkm7"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.503734 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.509543 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.510930 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511384 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511445 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/05f44398-4018-4ab3-a8c9-93f06d9b85cf-images\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511490 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-policies\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511515 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/349b87a8-e51b-4e41-a25d-c8eea29cca90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8f69k\" (UID: \"349b87a8-e51b-4e41-a25d-c8eea29cca90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511552 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c5efbd13-6819-4e2c-899f-1711650fbb48-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8x2nd\" (UID: \"c5efbd13-6819-4e2c-899f-1711650fbb48\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511593 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c085b695-1392-45eb-8fab-2813c719dfac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511665 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6afae383-3eb0-4bb9-8da0-6fa74ba1d09a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-57nxf\" (UID: \"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511699 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b1bb9823-6bb1-4e63-8729-efed71f79529-stats-auth\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511747 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6afae383-3eb0-4bb9-8da0-6fa74ba1d09a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-57nxf\" (UID: \"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511825 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511849 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c085b695-1392-45eb-8fab-2813c719dfac-service-ca-bundle\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511878 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c085b695-1392-45eb-8fab-2813c719dfac-serving-cert\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511944 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1bb9823-6bb1-4e63-8729-efed71f79529-metrics-certs\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.511988 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/765ad9da-3b9a-4647-a3ae-b68cee03e99b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hkwsg\" (UID: \"765ad9da-3b9a-4647-a3ae-b68cee03e99b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512052 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w5jj\" (UniqueName: \"kubernetes.io/projected/c085b695-1392-45eb-8fab-2813c719dfac-kube-api-access-9w5jj\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512087 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b1bb9823-6bb1-4e63-8729-efed71f79529-default-certificate\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512111 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmqmt\" (UniqueName: \"kubernetes.io/projected/05f44398-4018-4ab3-a8c9-93f06d9b85cf-kube-api-access-mmqmt\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512139 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1bb9823-6bb1-4e63-8729-efed71f79529-service-ca-bundle\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512200 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6d38885-8f9a-4840-832d-e831e5027583-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w4rlw\" (UID: \"c6d38885-8f9a-4840-832d-e831e5027583\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512223 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctrgk\" (UniqueName: \"kubernetes.io/projected/765ad9da-3b9a-4647-a3ae-b68cee03e99b-kube-api-access-ctrgk\") pod \"kube-storage-version-migrator-operator-b67b599dd-hkwsg\" (UID: \"765ad9da-3b9a-4647-a3ae-b68cee03e99b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512278 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-dir\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512334 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6d38885-8f9a-4840-832d-e831e5027583-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w4rlw\" (UID: \"c6d38885-8f9a-4840-832d-e831e5027583\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512418 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-dir\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512446 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdgfs\" (UniqueName: \"kubernetes.io/projected/c5efbd13-6819-4e2c-899f-1711650fbb48-kube-api-access-fdgfs\") pod \"machine-config-controller-84d6567774-8x2nd\" (UID: \"c5efbd13-6819-4e2c-899f-1711650fbb48\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512583 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512633 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512660 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6558\" (UniqueName: \"kubernetes.io/projected/f031a182-f40f-4369-83fa-51ba2f2ebd83-kube-api-access-f6558\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512719 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-serving-cert\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512924 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn6fs\" (UniqueName: \"kubernetes.io/projected/349b87a8-e51b-4e41-a25d-c8eea29cca90-kube-api-access-qn6fs\") pod \"openshift-apiserver-operator-796bbdcf4f-8f69k\" (UID: \"349b87a8-e51b-4e41-a25d-c8eea29cca90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.512960 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kvwx2"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.513719 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.513834 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-trusted-ca\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.513908 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67lw6\" (UniqueName: \"kubernetes.io/projected/e38a71c3-e5a2-4460-bc74-fc17bb897f67-kube-api-access-67lw6\") pod \"migrator-59844c95c7-m6l6j\" (UID: \"e38a71c3-e5a2-4460-bc74-fc17bb897f67\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.513964 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a917bbb-dee4-4c7d-852e-12f7be29bf33-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-w74x7\" (UID: \"2a917bbb-dee4-4c7d-852e-12f7be29bf33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.513988 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a917bbb-dee4-4c7d-852e-12f7be29bf33-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-w74x7\" (UID: \"2a917bbb-dee4-4c7d-852e-12f7be29bf33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514020 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-config\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514103 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-metrics-tls\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514322 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514361 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6afae383-3eb0-4bb9-8da0-6fa74ba1d09a-config\") pod \"kube-controller-manager-operator-78b949d7b-57nxf\" (UID: \"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514469 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514537 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514581 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514618 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7hsr\" (UniqueName: \"kubernetes.io/projected/2a917bbb-dee4-4c7d-852e-12f7be29bf33-kube-api-access-x7hsr\") pod \"openshift-controller-manager-operator-756b6f6bc6-w74x7\" (UID: \"2a917bbb-dee4-4c7d-852e-12f7be29bf33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514645 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514677 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/765ad9da-3b9a-4647-a3ae-b68cee03e99b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hkwsg\" (UID: \"765ad9da-3b9a-4647-a3ae-b68cee03e99b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514767 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12066056-05f8-4137-8977-f5287bc0d712-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8nrs6\" (UID: \"12066056-05f8-4137-8977-f5287bc0d712\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514851 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/349b87a8-e51b-4e41-a25d-c8eea29cca90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8f69k\" (UID: \"349b87a8-e51b-4e41-a25d-c8eea29cca90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.514917 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlxb7\" (UniqueName: \"kubernetes.io/projected/12066056-05f8-4137-8977-f5287bc0d712-kube-api-access-nlxb7\") pod \"multus-admission-controller-857f4d67dd-8nrs6\" (UID: \"12066056-05f8-4137-8977-f5287bc0d712\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.515203 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c085b695-1392-45eb-8fab-2813c719dfac-config\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.515566 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.515679 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-trusted-ca\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.516251 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9km9\" (UniqueName: \"kubernetes.io/projected/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-kube-api-access-g9km9\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.516120 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.516677 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.516923 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c5efbd13-6819-4e2c-899f-1711650fbb48-proxy-tls\") pod \"machine-config-controller-84d6567774-8x2nd\" (UID: \"c5efbd13-6819-4e2c-899f-1711650fbb48\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.516990 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05f44398-4018-4ab3-a8c9-93f06d9b85cf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.517022 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.517049 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5td4\" (UniqueName: \"kubernetes.io/projected/b1bb9823-6bb1-4e63-8729-efed71f79529-kube-api-access-r5td4\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.517079 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6d38885-8f9a-4840-832d-e831e5027583-config\") pod \"kube-apiserver-operator-766d6c64bb-w4rlw\" (UID: \"c6d38885-8f9a-4840-832d-e831e5027583\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.517142 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2gqg\" (UniqueName: \"kubernetes.io/projected/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-kube-api-access-j2gqg\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.517193 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05f44398-4018-4ab3-a8c9-93f06d9b85cf-proxy-tls\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.517326 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.518828 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5x65s"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.519594 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.519994 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.521356 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.522850 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.523697 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.528950 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8vdbw"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.529268 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.530930 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.531057 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.532981 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.534814 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.536080 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-rjlqg"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.537999 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.539750 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-8hxk6"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.539889 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.540573 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.540633 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.540690 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8hxk6" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.541144 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.542177 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.543296 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8nrs6"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.543554 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.544325 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-x5mfq"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.545430 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.546543 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.547851 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.549222 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.550057 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.550344 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8z4bc"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.552806 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8hxk6"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.554081 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-w49nv"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.555835 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-w49nv"] Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.556951 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.569746 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.590224 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.592534 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-policies\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.609812 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.615142 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618009 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1bb9823-6bb1-4e63-8729-efed71f79529-metrics-certs\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618055 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/765ad9da-3b9a-4647-a3ae-b68cee03e99b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hkwsg\" (UID: \"765ad9da-3b9a-4647-a3ae-b68cee03e99b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618090 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w5jj\" (UniqueName: \"kubernetes.io/projected/c085b695-1392-45eb-8fab-2813c719dfac-kube-api-access-9w5jj\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618130 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmqmt\" (UniqueName: \"kubernetes.io/projected/05f44398-4018-4ab3-a8c9-93f06d9b85cf-kube-api-access-mmqmt\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618155 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b1bb9823-6bb1-4e63-8729-efed71f79529-default-certificate\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618200 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1bb9823-6bb1-4e63-8729-efed71f79529-service-ca-bundle\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618236 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6d38885-8f9a-4840-832d-e831e5027583-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w4rlw\" (UID: \"c6d38885-8f9a-4840-832d-e831e5027583\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618272 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctrgk\" (UniqueName: \"kubernetes.io/projected/765ad9da-3b9a-4647-a3ae-b68cee03e99b-kube-api-access-ctrgk\") pod \"kube-storage-version-migrator-operator-b67b599dd-hkwsg\" (UID: \"765ad9da-3b9a-4647-a3ae-b68cee03e99b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618293 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6d38885-8f9a-4840-832d-e831e5027583-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w4rlw\" (UID: \"c6d38885-8f9a-4840-832d-e831e5027583\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618313 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdgfs\" (UniqueName: \"kubernetes.io/projected/c5efbd13-6819-4e2c-899f-1711650fbb48-kube-api-access-fdgfs\") pod \"machine-config-controller-84d6567774-8x2nd\" (UID: \"c5efbd13-6819-4e2c-899f-1711650fbb48\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618373 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-serving-cert\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618391 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn6fs\" (UniqueName: \"kubernetes.io/projected/349b87a8-e51b-4e41-a25d-c8eea29cca90-kube-api-access-qn6fs\") pod \"openshift-apiserver-operator-796bbdcf4f-8f69k\" (UID: \"349b87a8-e51b-4e41-a25d-c8eea29cca90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618432 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a917bbb-dee4-4c7d-852e-12f7be29bf33-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-w74x7\" (UID: \"2a917bbb-dee4-4c7d-852e-12f7be29bf33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618450 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a917bbb-dee4-4c7d-852e-12f7be29bf33-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-w74x7\" (UID: \"2a917bbb-dee4-4c7d-852e-12f7be29bf33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618468 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-trusted-ca\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618487 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67lw6\" (UniqueName: \"kubernetes.io/projected/e38a71c3-e5a2-4460-bc74-fc17bb897f67-kube-api-access-67lw6\") pod \"migrator-59844c95c7-m6l6j\" (UID: \"e38a71c3-e5a2-4460-bc74-fc17bb897f67\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618524 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-config\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618543 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-metrics-tls\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618562 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6afae383-3eb0-4bb9-8da0-6fa74ba1d09a-config\") pod \"kube-controller-manager-operator-78b949d7b-57nxf\" (UID: \"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618608 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618631 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7hsr\" (UniqueName: \"kubernetes.io/projected/2a917bbb-dee4-4c7d-852e-12f7be29bf33-kube-api-access-x7hsr\") pod \"openshift-controller-manager-operator-756b6f6bc6-w74x7\" (UID: \"2a917bbb-dee4-4c7d-852e-12f7be29bf33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618668 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/765ad9da-3b9a-4647-a3ae-b68cee03e99b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hkwsg\" (UID: \"765ad9da-3b9a-4647-a3ae-b68cee03e99b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618700 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12066056-05f8-4137-8977-f5287bc0d712-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8nrs6\" (UID: \"12066056-05f8-4137-8977-f5287bc0d712\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618720 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c085b695-1392-45eb-8fab-2813c719dfac-config\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618753 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/349b87a8-e51b-4e41-a25d-c8eea29cca90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8f69k\" (UID: \"349b87a8-e51b-4e41-a25d-c8eea29cca90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618772 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlxb7\" (UniqueName: \"kubernetes.io/projected/12066056-05f8-4137-8977-f5287bc0d712-kube-api-access-nlxb7\") pod \"multus-admission-controller-857f4d67dd-8nrs6\" (UID: \"12066056-05f8-4137-8977-f5287bc0d712\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618819 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-trusted-ca\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618922 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c5efbd13-6819-4e2c-899f-1711650fbb48-proxy-tls\") pod \"machine-config-controller-84d6567774-8x2nd\" (UID: \"c5efbd13-6819-4e2c-899f-1711650fbb48\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.618960 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9km9\" (UniqueName: \"kubernetes.io/projected/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-kube-api-access-g9km9\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.620348 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a917bbb-dee4-4c7d-852e-12f7be29bf33-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-w74x7\" (UID: \"2a917bbb-dee4-4c7d-852e-12f7be29bf33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.620446 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-config\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.620648 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-trusted-ca\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.620891 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1bb9823-6bb1-4e63-8729-efed71f79529-service-ca-bundle\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.621595 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c085b695-1392-45eb-8fab-2813c719dfac-config\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.621672 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05f44398-4018-4ab3-a8c9-93f06d9b85cf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.621739 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5td4\" (UniqueName: \"kubernetes.io/projected/b1bb9823-6bb1-4e63-8729-efed71f79529-kube-api-access-r5td4\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.621920 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6d38885-8f9a-4840-832d-e831e5027583-config\") pod \"kube-apiserver-operator-766d6c64bb-w4rlw\" (UID: \"c6d38885-8f9a-4840-832d-e831e5027583\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.622086 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2gqg\" (UniqueName: \"kubernetes.io/projected/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-kube-api-access-j2gqg\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.622400 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05f44398-4018-4ab3-a8c9-93f06d9b85cf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.622601 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1bb9823-6bb1-4e63-8729-efed71f79529-metrics-certs\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.622875 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05f44398-4018-4ab3-a8c9-93f06d9b85cf-proxy-tls\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.623743 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b1bb9823-6bb1-4e63-8729-efed71f79529-default-certificate\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.624267 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/05f44398-4018-4ab3-a8c9-93f06d9b85cf-images\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.624440 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c5efbd13-6819-4e2c-899f-1711650fbb48-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8x2nd\" (UID: \"c5efbd13-6819-4e2c-899f-1711650fbb48\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.624468 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/349b87a8-e51b-4e41-a25d-c8eea29cca90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8f69k\" (UID: \"349b87a8-e51b-4e41-a25d-c8eea29cca90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.625416 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c085b695-1392-45eb-8fab-2813c719dfac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.625837 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6afae383-3eb0-4bb9-8da0-6fa74ba1d09a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-57nxf\" (UID: \"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.625312 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a917bbb-dee4-4c7d-852e-12f7be29bf33-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-w74x7\" (UID: \"2a917bbb-dee4-4c7d-852e-12f7be29bf33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.625138 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-serving-cert\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.625380 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c5efbd13-6819-4e2c-899f-1711650fbb48-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8x2nd\" (UID: \"c5efbd13-6819-4e2c-899f-1711650fbb48\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.626007 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b1bb9823-6bb1-4e63-8729-efed71f79529-stats-auth\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.626245 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6afae383-3eb0-4bb9-8da0-6fa74ba1d09a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-57nxf\" (UID: \"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.626124 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c085b695-1392-45eb-8fab-2813c719dfac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.626343 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c085b695-1392-45eb-8fab-2813c719dfac-service-ca-bundle\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.626487 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c085b695-1392-45eb-8fab-2813c719dfac-serving-cert\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.626951 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c085b695-1392-45eb-8fab-2813c719dfac-service-ca-bundle\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.629932 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b1bb9823-6bb1-4e63-8729-efed71f79529-stats-auth\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.631077 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.633381 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c085b695-1392-45eb-8fab-2813c719dfac-serving-cert\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.650284 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.670366 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.685887 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6d38885-8f9a-4840-832d-e831e5027583-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w4rlw\" (UID: \"c6d38885-8f9a-4840-832d-e831e5027583\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.690329 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.692587 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6d38885-8f9a-4840-832d-e831e5027583-config\") pod \"kube-apiserver-operator-766d6c64bb-w4rlw\" (UID: \"c6d38885-8f9a-4840-832d-e831e5027583\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.709823 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.740172 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.748868 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-trusted-ca\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.752396 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.764724 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-metrics-tls\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.771464 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.791259 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.810253 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.832649 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.851162 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.861310 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6afae383-3eb0-4bb9-8da0-6fa74ba1d09a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-57nxf\" (UID: \"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.871552 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.880823 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6afae383-3eb0-4bb9-8da0-6fa74ba1d09a-config\") pod \"kube-controller-manager-operator-78b949d7b-57nxf\" (UID: \"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.891082 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.911245 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.931289 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.939168 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/349b87a8-e51b-4e41-a25d-c8eea29cca90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8f69k\" (UID: \"349b87a8-e51b-4e41-a25d-c8eea29cca90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.950355 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.951644 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/349b87a8-e51b-4e41-a25d-c8eea29cca90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8f69k\" (UID: \"349b87a8-e51b-4e41-a25d-c8eea29cca90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.970528 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.981944 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.981944 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.981975 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:11:07 crc kubenswrapper[4871]: I1007 22:11:07.990557 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.010687 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.030366 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.050721 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.070633 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.084136 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12066056-05f8-4137-8977-f5287bc0d712-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8nrs6\" (UID: \"12066056-05f8-4137-8977-f5287bc0d712\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.090771 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.095865 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c5efbd13-6819-4e2c-899f-1711650fbb48-proxy-tls\") pod \"machine-config-controller-84d6567774-8x2nd\" (UID: \"c5efbd13-6819-4e2c-899f-1711650fbb48\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.110909 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.131522 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.151347 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.170190 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.191181 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.195423 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/05f44398-4018-4ab3-a8c9-93f06d9b85cf-images\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.210916 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.231309 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.239956 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05f44398-4018-4ab3-a8c9-93f06d9b85cf-proxy-tls\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.250387 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.271141 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.291018 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.310463 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.351259 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.368941 4871 request.go:700] Waited for 1.017917733s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.371525 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.391966 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.410945 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.431014 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.450058 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/765ad9da-3b9a-4647-a3ae-b68cee03e99b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hkwsg\" (UID: \"765ad9da-3b9a-4647-a3ae-b68cee03e99b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.454020 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.460992 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/765ad9da-3b9a-4647-a3ae-b68cee03e99b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hkwsg\" (UID: \"765ad9da-3b9a-4647-a3ae-b68cee03e99b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.470977 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.510983 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.516186 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9pnm\" (UniqueName: \"kubernetes.io/projected/7b559a4d-6250-4309-b703-0d70e4652d12-kube-api-access-h9pnm\") pod \"machine-approver-56656f9798-np4z5\" (UID: \"7b559a4d-6250-4309-b703-0d70e4652d12\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.530588 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.551158 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.571329 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.610357 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.652092 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4p9r\" (UniqueName: \"kubernetes.io/projected/ecf86dbe-b51a-4228-a5ff-959dfe877387-kube-api-access-l4p9r\") pod \"apiserver-76f77b778f-5x65s\" (UID: \"ecf86dbe-b51a-4228-a5ff-959dfe877387\") " pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.671087 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.677587 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7hkb\" (UniqueName: \"kubernetes.io/projected/55ee5574-d8f1-48bd-83c3-7aa62e6366f1-kube-api-access-d7hkb\") pod \"openshift-config-operator-7777fb866f-9l8jh\" (UID: \"55ee5574-d8f1-48bd-83c3-7aa62e6366f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.688907 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.725163 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvrnv\" (UniqueName: \"kubernetes.io/projected/b4d26eda-8425-4a92-aeb1-418ea15e53fa-kube-api-access-cvrnv\") pod \"cluster-samples-operator-665b6dd947-j66lt\" (UID: \"b4d26eda-8425-4a92-aeb1-418ea15e53fa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.725942 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn2nm\" (UniqueName: \"kubernetes.io/projected/554a5462-ff56-4c07-b70b-0ff0f696f988-kube-api-access-mn2nm\") pod \"dns-operator-744455d44c-f5tkt\" (UID: \"554a5462-ff56-4c07-b70b-0ff0f696f988\") " pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.731759 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.736931 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.752204 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.775687 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.791351 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.833224 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.849428 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.852228 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.853357 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.871604 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.873541 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" event={"ID":"7b559a4d-6250-4309-b703-0d70e4652d12","Type":"ContainerStarted","Data":"0327f7b430ab87cfbb7a1f57478286b108483d4b210d7195b2beee4e35eec44f"} Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.915980 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbrc9\" (UniqueName: \"kubernetes.io/projected/0609d7b1-7443-4bcd-935e-ca2a61a359b7-kube-api-access-gbrc9\") pod \"controller-manager-879f6c89f-7lxmn\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.941457 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqqtl\" (UniqueName: \"kubernetes.io/projected/ff939f3b-6f07-4632-930b-0fa62db4d938-kube-api-access-gqqtl\") pod \"downloads-7954f5f757-hsm2n\" (UID: \"ff939f3b-6f07-4632-930b-0fa62db4d938\") " pod="openshift-console/downloads-7954f5f757-hsm2n" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.952284 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p2jk\" (UniqueName: \"kubernetes.io/projected/03ce8147-0273-421d-9263-3d4eb06bfcfe-kube-api-access-9p2jk\") pod \"route-controller-manager-6576b87f9c-48wk2\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.969658 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8znbq\" (UniqueName: \"kubernetes.io/projected/252cfed8-c315-49d0-911c-425cfb86fabb-kube-api-access-8znbq\") pod \"etcd-operator-b45778765-8vdbw\" (UID: \"252cfed8-c315-49d0-911c-425cfb86fabb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.976268 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5x65s"] Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.978334 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.981669 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:11:08 crc kubenswrapper[4871]: W1007 22:11:08.993919 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecf86dbe_b51a_4228_a5ff_959dfe877387.slice/crio-b2735dce62962e0c04ada5e636bfae21e2b3d4804299349589a2a7d60bdc8ff4 WatchSource:0}: Error finding container b2735dce62962e0c04ada5e636bfae21e2b3d4804299349589a2a7d60bdc8ff4: Status 404 returned error can't find the container with id b2735dce62962e0c04ada5e636bfae21e2b3d4804299349589a2a7d60bdc8ff4 Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.993996 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 07 22:11:08 crc kubenswrapper[4871]: I1007 22:11:08.994114 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c892w\" (UniqueName: \"kubernetes.io/projected/a6b7f864-334b-4d77-a8e4-bf62b614fb80-kube-api-access-c892w\") pod \"machine-api-operator-5694c8668f-2zz24\" (UID: \"a6b7f864-334b-4d77-a8e4-bf62b614fb80\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.000424 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.010393 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.011040 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.016969 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-hsm2n" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.031049 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.031070 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.053510 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.070893 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.077881 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.095125 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.109003 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh"] Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.113060 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.133344 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.154511 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.171600 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.209498 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkrmg\" (UniqueName: \"kubernetes.io/projected/f0312d2c-53d5-4916-9956-418acb3ff0b8-kube-api-access-wkrmg\") pod \"apiserver-7bbb656c7d-kbkgv\" (UID: \"f0312d2c-53d5-4916-9956-418acb3ff0b8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.213427 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.230703 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.232333 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.256302 4871 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.273075 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.312227 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.312496 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6558\" (UniqueName: \"kubernetes.io/projected/f031a182-f40f-4369-83fa-51ba2f2ebd83-kube-api-access-f6558\") pod \"oauth-openshift-558db77b4-clflh\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.331517 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.351286 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.368887 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.369479 4871 request.go:700] Waited for 1.828525234s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Ddefault-dockercfg-2llfx&limit=500&resourceVersion=0 Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.373091 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.392065 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.410103 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.416267 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.431361 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.452607 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2"] Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.453774 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.470761 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.481177 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2zz24"] Oct 07 22:11:09 crc kubenswrapper[4871]: W1007 22:11:09.484055 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03ce8147_0273_421d_9263_3d4eb06bfcfe.slice/crio-2213d8d2ac1334eb918233091e1b977693e60ddc8dfb859044fb09cbcb836666 WatchSource:0}: Error finding container 2213d8d2ac1334eb918233091e1b977693e60ddc8dfb859044fb09cbcb836666: Status 404 returned error can't find the container with id 2213d8d2ac1334eb918233091e1b977693e60ddc8dfb859044fb09cbcb836666 Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.491167 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.530556 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f5tkt"] Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.533729 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctrgk\" (UniqueName: \"kubernetes.io/projected/765ad9da-3b9a-4647-a3ae-b68cee03e99b-kube-api-access-ctrgk\") pod \"kube-storage-version-migrator-operator-b67b599dd-hkwsg\" (UID: \"765ad9da-3b9a-4647-a3ae-b68cee03e99b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:09 crc kubenswrapper[4871]: W1007 22:11:09.535335 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6b7f864_334b_4d77_a8e4_bf62b614fb80.slice/crio-51efa9c23e1e2808eadf5569f32ad93ead7b7b109738407c482d9bb168160400 WatchSource:0}: Error finding container 51efa9c23e1e2808eadf5569f32ad93ead7b7b109738407c482d9bb168160400: Status 404 returned error can't find the container with id 51efa9c23e1e2808eadf5569f32ad93ead7b7b109738407c482d9bb168160400 Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.542814 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-hsm2n"] Oct 07 22:11:09 crc kubenswrapper[4871]: W1007 22:11:09.550855 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod554a5462_ff56_4c07_b70b_0ff0f696f988.slice/crio-5fcfd7566df31698daa3bd7805f5c8ef27866740c2fc0120b99abaf8b2221298 WatchSource:0}: Error finding container 5fcfd7566df31698daa3bd7805f5c8ef27866740c2fc0120b99abaf8b2221298: Status 404 returned error can't find the container with id 5fcfd7566df31698daa3bd7805f5c8ef27866740c2fc0120b99abaf8b2221298 Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.568400 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w5jj\" (UniqueName: \"kubernetes.io/projected/c085b695-1392-45eb-8fab-2813c719dfac-kube-api-access-9w5jj\") pod \"authentication-operator-69f744f599-5gxf8\" (UID: \"c085b695-1392-45eb-8fab-2813c719dfac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.577094 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt"] Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.583820 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmqmt\" (UniqueName: \"kubernetes.io/projected/05f44398-4018-4ab3-a8c9-93f06d9b85cf-kube-api-access-mmqmt\") pod \"machine-config-operator-74547568cd-2cq6r\" (UID: \"05f44398-4018-4ab3-a8c9-93f06d9b85cf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.588447 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.614394 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-clflh"] Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.615883 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7hsr\" (UniqueName: \"kubernetes.io/projected/2a917bbb-dee4-4c7d-852e-12f7be29bf33-kube-api-access-x7hsr\") pod \"openshift-controller-manager-operator-756b6f6bc6-w74x7\" (UID: \"2a917bbb-dee4-4c7d-852e-12f7be29bf33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.628048 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8vdbw"] Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.629922 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7lxmn"] Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.629987 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.637657 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67lw6\" (UniqueName: \"kubernetes.io/projected/e38a71c3-e5a2-4460-bc74-fc17bb897f67-kube-api-access-67lw6\") pod \"migrator-59844c95c7-m6l6j\" (UID: \"e38a71c3-e5a2-4460-bc74-fc17bb897f67\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.642596 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" Oct 07 22:11:09 crc kubenswrapper[4871]: W1007 22:11:09.646495 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf031a182_f40f_4369_83fa_51ba2f2ebd83.slice/crio-5a090c0ebb6c4cbaee00164f58f606ede634f61a25768c1569345fd534eae44a WatchSource:0}: Error finding container 5a090c0ebb6c4cbaee00164f58f606ede634f61a25768c1569345fd534eae44a: Status 404 returned error can't find the container with id 5a090c0ebb6c4cbaee00164f58f606ede634f61a25768c1569345fd534eae44a Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.650858 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlxb7\" (UniqueName: \"kubernetes.io/projected/12066056-05f8-4137-8977-f5287bc0d712-kube-api-access-nlxb7\") pod \"multus-admission-controller-857f4d67dd-8nrs6\" (UID: \"12066056-05f8-4137-8977-f5287bc0d712\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.677144 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdgfs\" (UniqueName: \"kubernetes.io/projected/c5efbd13-6819-4e2c-899f-1711650fbb48-kube-api-access-fdgfs\") pod \"machine-config-controller-84d6567774-8x2nd\" (UID: \"c5efbd13-6819-4e2c-899f-1711650fbb48\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.689648 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9km9\" (UniqueName: \"kubernetes.io/projected/a0825d5d-ec58-4a42-85e6-7678c5a6a5bc-kube-api-access-g9km9\") pod \"ingress-operator-5b745b69d9-2f6w8\" (UID: \"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.696899 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv"] Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.705442 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.707033 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn6fs\" (UniqueName: \"kubernetes.io/projected/349b87a8-e51b-4e41-a25d-c8eea29cca90-kube-api-access-qn6fs\") pod \"openshift-apiserver-operator-796bbdcf4f-8f69k\" (UID: \"349b87a8-e51b-4e41-a25d-c8eea29cca90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.710636 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.718195 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.726211 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.729983 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6d38885-8f9a-4840-832d-e831e5027583-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w4rlw\" (UID: \"c6d38885-8f9a-4840-832d-e831e5027583\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.743632 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.750122 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5td4\" (UniqueName: \"kubernetes.io/projected/b1bb9823-6bb1-4e63-8729-efed71f79529-kube-api-access-r5td4\") pod \"router-default-5444994796-r8db4\" (UID: \"b1bb9823-6bb1-4e63-8729-efed71f79529\") " pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.766601 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2gqg\" (UniqueName: \"kubernetes.io/projected/7e7841e7-85ae-4d41-acf4-a4603f33d6d4-kube-api-access-j2gqg\") pod \"console-operator-58897d9998-bwsvx\" (UID: \"7e7841e7-85ae-4d41-acf4-a4603f33d6d4\") " pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.788434 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6afae383-3eb0-4bb9-8da0-6fa74ba1d09a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-57nxf\" (UID: \"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.790113 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.810704 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.831304 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.852497 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.893352 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" event={"ID":"f031a182-f40f-4369-83fa-51ba2f2ebd83","Type":"ContainerStarted","Data":"5a090c0ebb6c4cbaee00164f58f606ede634f61a25768c1569345fd534eae44a"} Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.896516 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" event={"ID":"b4d26eda-8425-4a92-aeb1-418ea15e53fa","Type":"ContainerStarted","Data":"b4cfff6abdebc73b3f08c49f95f8840355e4e8613f8554b47b8061c7c9c9f11b"} Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.896574 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" event={"ID":"b4d26eda-8425-4a92-aeb1-418ea15e53fa","Type":"ContainerStarted","Data":"d2e5d1f6b4cd22f9c5e91e405b072a63949795c3d3cf9c115cd84b533b851106"} Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.897427 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" event={"ID":"554a5462-ff56-4c07-b70b-0ff0f696f988","Type":"ContainerStarted","Data":"5fcfd7566df31698daa3bd7805f5c8ef27866740c2fc0120b99abaf8b2221298"} Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.906308 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" event={"ID":"03ce8147-0273-421d-9263-3d4eb06bfcfe","Type":"ContainerStarted","Data":"0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796"} Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.906352 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" event={"ID":"03ce8147-0273-421d-9263-3d4eb06bfcfe","Type":"ContainerStarted","Data":"2213d8d2ac1334eb918233091e1b977693e60ddc8dfb859044fb09cbcb836666"} Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.907083 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.907140 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5gxf8"] Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.911296 4871 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-48wk2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.911342 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" podUID="03ce8147-0273-421d-9263-3d4eb06bfcfe" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.911521 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.930518 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.958986 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.970190 4871 generic.go:334] "Generic (PLEG): container finished" podID="55ee5574-d8f1-48bd-83c3-7aa62e6366f1" containerID="23b570e338534c0c77964b65d701d166abd332071b40f451c1339c60c33a2a98" exitCode=0 Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.971830 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" event={"ID":"55ee5574-d8f1-48bd-83c3-7aa62e6366f1","Type":"ContainerDied","Data":"23b570e338534c0c77964b65d701d166abd332071b40f451c1339c60c33a2a98"} Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.971903 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" event={"ID":"55ee5574-d8f1-48bd-83c3-7aa62e6366f1","Type":"ContainerStarted","Data":"d12cc8c11cbb4ccbd2a3f9ff67faf7e38bd0dc9448331265eace31f0f66e11be"} Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.972300 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977244 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3689423d-c1f1-46c8-9ea5-165147a8a286-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977372 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-oauth-serving-cert\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977406 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kqcm9\" (UID: \"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977430 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-config\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977448 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvhl7\" (UniqueName: \"kubernetes.io/projected/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-kube-api-access-nvhl7\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977491 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3689423d-c1f1-46c8-9ea5-165147a8a286-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977510 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977528 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-serving-cert\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977577 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kqcm9\" (UID: \"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977632 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-certificates\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977655 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m79z\" (UniqueName: \"kubernetes.io/projected/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-kube-api-access-2m79z\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977671 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-tls\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977714 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kqcm9\" (UID: \"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977748 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-bound-sa-token\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977850 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgrft\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-kube-api-access-rgrft\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.977867 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-oauth-config\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.978029 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.978069 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.978154 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-trusted-ca\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.978182 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-service-ca\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.978214 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.978237 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-trusted-ca-bundle\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.978268 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/18a8f350-7724-4302-8663-499480b9af6e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cc8tr\" (UID: \"18a8f350-7724-4302-8663-499480b9af6e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.979428 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctb5q\" (UniqueName: \"kubernetes.io/projected/18a8f350-7724-4302-8663-499480b9af6e-kube-api-access-ctb5q\") pod \"control-plane-machine-set-operator-78cbb6b69f-cc8tr\" (UID: \"18a8f350-7724-4302-8663-499480b9af6e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.979496 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" Oct 07 22:11:09 crc kubenswrapper[4871]: E1007 22:11:09.981015 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:10.480989977 +0000 UTC m=+144.283688250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.985358 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" Oct 07 22:11:09 crc kubenswrapper[4871]: I1007 22:11:09.992936 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.021265 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" event={"ID":"f0312d2c-53d5-4916-9956-418acb3ff0b8","Type":"ContainerStarted","Data":"e05b648d779a61c700753d3cdf815271274f1f7a5218d61dd93c05a0b2eea8fd"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.023147 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7"] Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.046133 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.055595 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" event={"ID":"7b559a4d-6250-4309-b703-0d70e4652d12","Type":"ContainerStarted","Data":"5636ae5f792631e2d5f96b19cda2f3e03e70f21bd0334917c7e317772271b871"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.055671 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" event={"ID":"7b559a4d-6250-4309-b703-0d70e4652d12","Type":"ContainerStarted","Data":"cb8637ae4da40296a9c8cb56df532765095cd6119910686331726beee5f8acee"} Oct 07 22:11:10 crc kubenswrapper[4871]: W1007 22:11:10.073155 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a917bbb_dee4_4c7d_852e_12f7be29bf33.slice/crio-9da70dc42dba1bed8eed1dcfdb1baca67ba4c1c3ccb3adc496d333ca4e5dc80d WatchSource:0}: Error finding container 9da70dc42dba1bed8eed1dcfdb1baca67ba4c1c3ccb3adc496d333ca4e5dc80d: Status 404 returned error can't find the container with id 9da70dc42dba1bed8eed1dcfdb1baca67ba4c1c3ccb3adc496d333ca4e5dc80d Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082134 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082372 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-config\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082409 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srjbm\" (UniqueName: \"kubernetes.io/projected/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-kube-api-access-srjbm\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082432 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvhl7\" (UniqueName: \"kubernetes.io/projected/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-kube-api-access-nvhl7\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082453 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a916502-6bc9-48ba-906e-831e6fe85e3b-secret-volume\") pod \"collect-profiles-29331240-6x9cw\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082477 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvz6c\" (UniqueName: \"kubernetes.io/projected/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-kube-api-access-mvz6c\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082502 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-mountpoint-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082523 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3689423d-c1f1-46c8-9ea5-165147a8a286-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082546 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082573 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f6d85de4-a099-42e6-bb04-7016ed806e86-node-bootstrap-token\") pod \"machine-config-server-rjlqg\" (UID: \"f6d85de4-a099-42e6-bb04-7016ed806e86\") " pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082598 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-serving-cert\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082621 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp98p\" (UniqueName: \"kubernetes.io/projected/7bccb117-9b78-4a2a-b2fb-34952249e63c-kube-api-access-lp98p\") pod \"dns-default-w49nv\" (UID: \"7bccb117-9b78-4a2a-b2fb-34952249e63c\") " pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082690 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kqcm9\" (UID: \"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.082718 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/348abc9a-ac32-4c1a-8af5-16a95155591c-profile-collector-cert\") pod \"catalog-operator-68c6474976-zgsc2\" (UID: \"348abc9a-ac32-4c1a-8af5-16a95155591c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.083920 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:10.583892108 +0000 UTC m=+144.386590371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.084395 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-socket-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.084506 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-certificates\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.084615 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kvwx2\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.084660 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f6d85de4-a099-42e6-bb04-7016ed806e86-certs\") pod \"machine-config-server-rjlqg\" (UID: \"f6d85de4-a099-42e6-bb04-7016ed806e86\") " pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.085187 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e9ec82f3-8972-4d39-b35d-3f2901beb829-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5722h\" (UID: \"e9ec82f3-8972-4d39-b35d-3f2901beb829\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.085218 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kvwx2\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.085258 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m79z\" (UniqueName: \"kubernetes.io/projected/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-kube-api-access-2m79z\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.085291 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prp6p\" (UniqueName: \"kubernetes.io/projected/496b361b-01ac-48a6-ab37-c7bea109017d-kube-api-access-prp6p\") pod \"package-server-manager-789f6589d5-4p54k\" (UID: \"496b361b-01ac-48a6-ab37-c7bea109017d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.085320 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kqcm9\" (UID: \"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.086395 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-config\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.087270 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-tls\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.087525 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-certificates\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.088389 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cl8z\" (UniqueName: \"kubernetes.io/projected/9d97e32b-02b4-4435-bd25-81174b3e0da9-kube-api-access-8cl8z\") pod \"marketplace-operator-79b997595-kvwx2\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.088463 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kqcm9\" (UID: \"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.089288 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3689423d-c1f1-46c8-9ea5-165147a8a286-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.089304 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-csi-data-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.089372 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-plugins-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.089417 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-bound-sa-token\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.089505 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/fd949a6d-e0ef-4982-ad47-ea206a6faa58-signing-key\") pod \"service-ca-9c57cc56f-x5mfq\" (UID: \"fd949a6d-e0ef-4982-ad47-ea206a6faa58\") " pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.089527 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/496b361b-01ac-48a6-ab37-c7bea109017d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4p54k\" (UID: \"496b361b-01ac-48a6-ab37-c7bea109017d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.089547 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-registration-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.089598 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkncv\" (UniqueName: \"kubernetes.io/projected/fd949a6d-e0ef-4982-ad47-ea206a6faa58-kube-api-access-lkncv\") pod \"service-ca-9c57cc56f-x5mfq\" (UID: \"fd949a6d-e0ef-4982-ad47-ea206a6faa58\") " pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.096589 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-tls\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.096849 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq5cq\" (UniqueName: \"kubernetes.io/projected/348abc9a-ac32-4c1a-8af5-16a95155591c-kube-api-access-tq5cq\") pod \"catalog-operator-68c6474976-zgsc2\" (UID: \"348abc9a-ac32-4c1a-8af5-16a95155591c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.097910 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-serving-cert\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.100723 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgrft\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-kube-api-access-rgrft\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.100811 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-oauth-config\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.100935 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-apiservice-cert\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.107533 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.107599 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.107746 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30ac8190-e410-439f-9d15-95a62b9096b6-cert\") pod \"ingress-canary-8hxk6\" (UID: \"30ac8190-e410-439f-9d15-95a62b9096b6\") " pod="openshift-ingress-canary/ingress-canary-8hxk6" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.107816 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e9ec82f3-8972-4d39-b35d-3f2901beb829-srv-cert\") pod \"olm-operator-6b444d44fb-5722h\" (UID: \"e9ec82f3-8972-4d39-b35d-3f2901beb829\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.107853 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-tmpfs\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.108028 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-oauth-config\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.108069 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57a0865b-bdb2-4a39-a750-e1b5e77b2e64-serving-cert\") pod \"service-ca-operator-777779d784-cv9kf\" (UID: \"57a0865b-bdb2-4a39-a750-e1b5e77b2e64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.108553 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:10.608532334 +0000 UTC m=+144.411230407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.108737 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz2kr\" (UniqueName: \"kubernetes.io/projected/1a916502-6bc9-48ba-906e-831e6fe85e3b-kube-api-access-bz2kr\") pod \"collect-profiles-29331240-6x9cw\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.109275 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-trusted-ca\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.109353 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-service-ca\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.109391 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a916502-6bc9-48ba-906e-831e6fe85e3b-config-volume\") pod \"collect-profiles-29331240-6x9cw\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.109825 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.109930 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rftmz\" (UniqueName: \"kubernetes.io/projected/f6d85de4-a099-42e6-bb04-7016ed806e86-kube-api-access-rftmz\") pod \"machine-config-server-rjlqg\" (UID: \"f6d85de4-a099-42e6-bb04-7016ed806e86\") " pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.110341 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.110632 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qctsz\" (UniqueName: \"kubernetes.io/projected/57a0865b-bdb2-4a39-a750-e1b5e77b2e64-kube-api-access-qctsz\") pod \"service-ca-operator-777779d784-cv9kf\" (UID: \"57a0865b-bdb2-4a39-a750-e1b5e77b2e64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.110709 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-trusted-ca-bundle\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.110749 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7bccb117-9b78-4a2a-b2fb-34952249e63c-config-volume\") pod \"dns-default-w49nv\" (UID: \"7bccb117-9b78-4a2a-b2fb-34952249e63c\") " pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.111170 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/18a8f350-7724-4302-8663-499480b9af6e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cc8tr\" (UID: \"18a8f350-7724-4302-8663-499480b9af6e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.111324 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-webhook-cert\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.111373 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd24m\" (UniqueName: \"kubernetes.io/projected/e9ec82f3-8972-4d39-b35d-3f2901beb829-kube-api-access-nd24m\") pod \"olm-operator-6b444d44fb-5722h\" (UID: \"e9ec82f3-8972-4d39-b35d-3f2901beb829\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.112117 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/fd949a6d-e0ef-4982-ad47-ea206a6faa58-signing-cabundle\") pod \"service-ca-9c57cc56f-x5mfq\" (UID: \"fd949a6d-e0ef-4982-ad47-ea206a6faa58\") " pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.112213 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctb5q\" (UniqueName: \"kubernetes.io/projected/18a8f350-7724-4302-8663-499480b9af6e-kube-api-access-ctb5q\") pod \"control-plane-machine-set-operator-78cbb6b69f-cc8tr\" (UID: \"18a8f350-7724-4302-8663-499480b9af6e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.113350 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7bccb117-9b78-4a2a-b2fb-34952249e63c-metrics-tls\") pod \"dns-default-w49nv\" (UID: \"7bccb117-9b78-4a2a-b2fb-34952249e63c\") " pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.113858 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-trusted-ca\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.114252 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3689423d-c1f1-46c8-9ea5-165147a8a286-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.115659 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-service-ca\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.116188 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f95pf\" (UniqueName: \"kubernetes.io/projected/30ac8190-e410-439f-9d15-95a62b9096b6-kube-api-access-f95pf\") pod \"ingress-canary-8hxk6\" (UID: \"30ac8190-e410-439f-9d15-95a62b9096b6\") " pod="openshift-ingress-canary/ingress-canary-8hxk6" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.116664 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-oauth-serving-cert\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.116716 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a0865b-bdb2-4a39-a750-e1b5e77b2e64-config\") pod \"service-ca-operator-777779d784-cv9kf\" (UID: \"57a0865b-bdb2-4a39-a750-e1b5e77b2e64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.117215 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/348abc9a-ac32-4c1a-8af5-16a95155591c-srv-cert\") pod \"catalog-operator-68c6474976-zgsc2\" (UID: \"348abc9a-ac32-4c1a-8af5-16a95155591c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.117354 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.118120 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kqcm9\" (UID: \"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.119514 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3689423d-c1f1-46c8-9ea5-165147a8a286-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.120510 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/18a8f350-7724-4302-8663-499480b9af6e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cc8tr\" (UID: \"18a8f350-7724-4302-8663-499480b9af6e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.121126 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-oauth-serving-cert\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.123655 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-trusted-ca-bundle\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.129443 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kqcm9\" (UID: \"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.133248 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hsm2n" event={"ID":"ff939f3b-6f07-4632-930b-0fa62db4d938","Type":"ContainerStarted","Data":"7656e246f6794ec54d80ae16d8b55241517217e097ae1e1ac4f425f02631f845"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.133463 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hsm2n" event={"ID":"ff939f3b-6f07-4632-930b-0fa62db4d938","Type":"ContainerStarted","Data":"054ff31dec8d6df620d1ff4cbf14ff861cadbffcdaedfe6928b432e63c9409f5"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.134216 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-hsm2n" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.141551 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.147040 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" event={"ID":"0609d7b1-7443-4bcd-935e-ca2a61a359b7","Type":"ContainerStarted","Data":"fb3cecb1a77b5b1156ad94c9c416959ab89ab44f6cd9363393761286eb16efea"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.148098 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.150614 4871 patch_prober.go:28] interesting pod/downloads-7954f5f757-hsm2n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.150681 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hsm2n" podUID="ff939f3b-6f07-4632-930b-0fa62db4d938" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.152431 4871 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-7lxmn container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.152497 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" podUID="0609d7b1-7443-4bcd-935e-ca2a61a359b7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.163088 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvhl7\" (UniqueName: \"kubernetes.io/projected/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-kube-api-access-nvhl7\") pod \"console-f9d7485db-tw5sf\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.163688 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" event={"ID":"a6b7f864-334b-4d77-a8e4-bf62b614fb80","Type":"ContainerStarted","Data":"27d4af6d0c5d397efda032f593ae0ca31f79e88d149baf4511e7b91b719e1b51"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.163762 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" event={"ID":"a6b7f864-334b-4d77-a8e4-bf62b614fb80","Type":"ContainerStarted","Data":"51efa9c23e1e2808eadf5569f32ad93ead7b7b109738407c482d9bb168160400"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.167088 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" event={"ID":"252cfed8-c315-49d0-911c-425cfb86fabb","Type":"ContainerStarted","Data":"dc24314f38cee868465fc44da9722e93f2d748ae606973ee69d02a58e6d45849"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.167673 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kqcm9\" (UID: \"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.172461 4871 generic.go:334] "Generic (PLEG): container finished" podID="ecf86dbe-b51a-4228-a5ff-959dfe877387" containerID="f02f79a30d83ad545e2a221fcc8429abe75aae98f3e69687433122eceb06eb28" exitCode=0 Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.172527 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" event={"ID":"ecf86dbe-b51a-4228-a5ff-959dfe877387","Type":"ContainerDied","Data":"f02f79a30d83ad545e2a221fcc8429abe75aae98f3e69687433122eceb06eb28"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.172564 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" event={"ID":"ecf86dbe-b51a-4228-a5ff-959dfe877387","Type":"ContainerStarted","Data":"b2735dce62962e0c04ada5e636bfae21e2b3d4804299349589a2a7d60bdc8ff4"} Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.193683 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m79z\" (UniqueName: \"kubernetes.io/projected/159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0-kube-api-access-2m79z\") pod \"cluster-image-registry-operator-dc59b4c8b-mvq76\" (UID: \"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:10 crc kubenswrapper[4871]: W1007 22:11:10.202414 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1bb9823_6bb1_4e63_8729_efed71f79529.slice/crio-197963a6e4b40124baa3dd5709aac9ec9eb56deeb0846e81a05a76502be5d0f1 WatchSource:0}: Error finding container 197963a6e4b40124baa3dd5709aac9ec9eb56deeb0846e81a05a76502be5d0f1: Status 404 returned error can't find the container with id 197963a6e4b40124baa3dd5709aac9ec9eb56deeb0846e81a05a76502be5d0f1 Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.214643 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-bound-sa-token\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.219756 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.219998 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:10.719951542 +0000 UTC m=+144.522649615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220063 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srjbm\" (UniqueName: \"kubernetes.io/projected/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-kube-api-access-srjbm\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220129 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvz6c\" (UniqueName: \"kubernetes.io/projected/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-kube-api-access-mvz6c\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220157 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-mountpoint-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220184 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a916502-6bc9-48ba-906e-831e6fe85e3b-secret-volume\") pod \"collect-profiles-29331240-6x9cw\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220206 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f6d85de4-a099-42e6-bb04-7016ed806e86-node-bootstrap-token\") pod \"machine-config-server-rjlqg\" (UID: \"f6d85de4-a099-42e6-bb04-7016ed806e86\") " pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220263 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp98p\" (UniqueName: \"kubernetes.io/projected/7bccb117-9b78-4a2a-b2fb-34952249e63c-kube-api-access-lp98p\") pod \"dns-default-w49nv\" (UID: \"7bccb117-9b78-4a2a-b2fb-34952249e63c\") " pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220317 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/348abc9a-ac32-4c1a-8af5-16a95155591c-profile-collector-cert\") pod \"catalog-operator-68c6474976-zgsc2\" (UID: \"348abc9a-ac32-4c1a-8af5-16a95155591c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220353 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-socket-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220390 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kvwx2\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220419 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f6d85de4-a099-42e6-bb04-7016ed806e86-certs\") pod \"machine-config-server-rjlqg\" (UID: \"f6d85de4-a099-42e6-bb04-7016ed806e86\") " pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220449 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e9ec82f3-8972-4d39-b35d-3f2901beb829-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5722h\" (UID: \"e9ec82f3-8972-4d39-b35d-3f2901beb829\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.220879 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-mountpoint-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.221476 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-socket-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.225662 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kvwx2\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.225743 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prp6p\" (UniqueName: \"kubernetes.io/projected/496b361b-01ac-48a6-ab37-c7bea109017d-kube-api-access-prp6p\") pod \"package-server-manager-789f6589d5-4p54k\" (UID: \"496b361b-01ac-48a6-ab37-c7bea109017d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.225858 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cl8z\" (UniqueName: \"kubernetes.io/projected/9d97e32b-02b4-4435-bd25-81174b3e0da9-kube-api-access-8cl8z\") pod \"marketplace-operator-79b997595-kvwx2\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.225900 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-csi-data-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.225934 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-plugins-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.225972 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/fd949a6d-e0ef-4982-ad47-ea206a6faa58-signing-key\") pod \"service-ca-9c57cc56f-x5mfq\" (UID: \"fd949a6d-e0ef-4982-ad47-ea206a6faa58\") " pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.225997 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/496b361b-01ac-48a6-ab37-c7bea109017d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4p54k\" (UID: \"496b361b-01ac-48a6-ab37-c7bea109017d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226024 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-registration-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226053 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkncv\" (UniqueName: \"kubernetes.io/projected/fd949a6d-e0ef-4982-ad47-ea206a6faa58-kube-api-access-lkncv\") pod \"service-ca-9c57cc56f-x5mfq\" (UID: \"fd949a6d-e0ef-4982-ad47-ea206a6faa58\") " pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226084 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq5cq\" (UniqueName: \"kubernetes.io/projected/348abc9a-ac32-4c1a-8af5-16a95155591c-kube-api-access-tq5cq\") pod \"catalog-operator-68c6474976-zgsc2\" (UID: \"348abc9a-ac32-4c1a-8af5-16a95155591c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226341 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-plugins-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226631 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-apiservice-cert\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226738 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-csi-data-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226767 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-registration-dir\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226735 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226844 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30ac8190-e410-439f-9d15-95a62b9096b6-cert\") pod \"ingress-canary-8hxk6\" (UID: \"30ac8190-e410-439f-9d15-95a62b9096b6\") " pod="openshift-ingress-canary/ingress-canary-8hxk6" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.226871 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e9ec82f3-8972-4d39-b35d-3f2901beb829-srv-cert\") pod \"olm-operator-6b444d44fb-5722h\" (UID: \"e9ec82f3-8972-4d39-b35d-3f2901beb829\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.227405 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kvwx2\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.227461 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/348abc9a-ac32-4c1a-8af5-16a95155591c-profile-collector-cert\") pod \"catalog-operator-68c6474976-zgsc2\" (UID: \"348abc9a-ac32-4c1a-8af5-16a95155591c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.227776 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:10.72775319 +0000 UTC m=+144.530451253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.227887 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-tmpfs\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.231342 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57a0865b-bdb2-4a39-a750-e1b5e77b2e64-serving-cert\") pod \"service-ca-operator-777779d784-cv9kf\" (UID: \"57a0865b-bdb2-4a39-a750-e1b5e77b2e64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.231513 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz2kr\" (UniqueName: \"kubernetes.io/projected/1a916502-6bc9-48ba-906e-831e6fe85e3b-kube-api-access-bz2kr\") pod \"collect-profiles-29331240-6x9cw\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.231935 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a916502-6bc9-48ba-906e-831e6fe85e3b-secret-volume\") pod \"collect-profiles-29331240-6x9cw\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.231951 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a916502-6bc9-48ba-906e-831e6fe85e3b-config-volume\") pod \"collect-profiles-29331240-6x9cw\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.232136 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rftmz\" (UniqueName: \"kubernetes.io/projected/f6d85de4-a099-42e6-bb04-7016ed806e86-kube-api-access-rftmz\") pod \"machine-config-server-rjlqg\" (UID: \"f6d85de4-a099-42e6-bb04-7016ed806e86\") " pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.233090 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgrft\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-kube-api-access-rgrft\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.233759 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f6d85de4-a099-42e6-bb04-7016ed806e86-node-bootstrap-token\") pod \"machine-config-server-rjlqg\" (UID: \"f6d85de4-a099-42e6-bb04-7016ed806e86\") " pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.235140 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a916502-6bc9-48ba-906e-831e6fe85e3b-config-volume\") pod \"collect-profiles-29331240-6x9cw\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.238720 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-tmpfs\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.238920 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qctsz\" (UniqueName: \"kubernetes.io/projected/57a0865b-bdb2-4a39-a750-e1b5e77b2e64-kube-api-access-qctsz\") pod \"service-ca-operator-777779d784-cv9kf\" (UID: \"57a0865b-bdb2-4a39-a750-e1b5e77b2e64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.239574 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57a0865b-bdb2-4a39-a750-e1b5e77b2e64-serving-cert\") pod \"service-ca-operator-777779d784-cv9kf\" (UID: \"57a0865b-bdb2-4a39-a750-e1b5e77b2e64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.239589 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7bccb117-9b78-4a2a-b2fb-34952249e63c-config-volume\") pod \"dns-default-w49nv\" (UID: \"7bccb117-9b78-4a2a-b2fb-34952249e63c\") " pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.239643 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-webhook-cert\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.239679 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd24m\" (UniqueName: \"kubernetes.io/projected/e9ec82f3-8972-4d39-b35d-3f2901beb829-kube-api-access-nd24m\") pod \"olm-operator-6b444d44fb-5722h\" (UID: \"e9ec82f3-8972-4d39-b35d-3f2901beb829\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.239703 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/fd949a6d-e0ef-4982-ad47-ea206a6faa58-signing-cabundle\") pod \"service-ca-9c57cc56f-x5mfq\" (UID: \"fd949a6d-e0ef-4982-ad47-ea206a6faa58\") " pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.239757 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7bccb117-9b78-4a2a-b2fb-34952249e63c-metrics-tls\") pod \"dns-default-w49nv\" (UID: \"7bccb117-9b78-4a2a-b2fb-34952249e63c\") " pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.239916 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f95pf\" (UniqueName: \"kubernetes.io/projected/30ac8190-e410-439f-9d15-95a62b9096b6-kube-api-access-f95pf\") pod \"ingress-canary-8hxk6\" (UID: \"30ac8190-e410-439f-9d15-95a62b9096b6\") " pod="openshift-ingress-canary/ingress-canary-8hxk6" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.239975 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a0865b-bdb2-4a39-a750-e1b5e77b2e64-config\") pod \"service-ca-operator-777779d784-cv9kf\" (UID: \"57a0865b-bdb2-4a39-a750-e1b5e77b2e64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.240001 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/348abc9a-ac32-4c1a-8af5-16a95155591c-srv-cert\") pod \"catalog-operator-68c6474976-zgsc2\" (UID: \"348abc9a-ac32-4c1a-8af5-16a95155591c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.242438 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kvwx2\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.242630 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f6d85de4-a099-42e6-bb04-7016ed806e86-certs\") pod \"machine-config-server-rjlqg\" (UID: \"f6d85de4-a099-42e6-bb04-7016ed806e86\") " pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.243447 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/fd949a6d-e0ef-4982-ad47-ea206a6faa58-signing-key\") pod \"service-ca-9c57cc56f-x5mfq\" (UID: \"fd949a6d-e0ef-4982-ad47-ea206a6faa58\") " pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.243550 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-apiservice-cert\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.244238 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e9ec82f3-8972-4d39-b35d-3f2901beb829-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5722h\" (UID: \"e9ec82f3-8972-4d39-b35d-3f2901beb829\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.244850 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-webhook-cert\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.246904 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30ac8190-e410-439f-9d15-95a62b9096b6-cert\") pod \"ingress-canary-8hxk6\" (UID: \"30ac8190-e410-439f-9d15-95a62b9096b6\") " pod="openshift-ingress-canary/ingress-canary-8hxk6" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.247022 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.247370 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e9ec82f3-8972-4d39-b35d-3f2901beb829-srv-cert\") pod \"olm-operator-6b444d44fb-5722h\" (UID: \"e9ec82f3-8972-4d39-b35d-3f2901beb829\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.247701 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7bccb117-9b78-4a2a-b2fb-34952249e63c-metrics-tls\") pod \"dns-default-w49nv\" (UID: \"7bccb117-9b78-4a2a-b2fb-34952249e63c\") " pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.247959 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a0865b-bdb2-4a39-a750-e1b5e77b2e64-config\") pod \"service-ca-operator-777779d784-cv9kf\" (UID: \"57a0865b-bdb2-4a39-a750-e1b5e77b2e64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.250921 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/496b361b-01ac-48a6-ab37-c7bea109017d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4p54k\" (UID: \"496b361b-01ac-48a6-ab37-c7bea109017d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.251049 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/fd949a6d-e0ef-4982-ad47-ea206a6faa58-signing-cabundle\") pod \"service-ca-9c57cc56f-x5mfq\" (UID: \"fd949a6d-e0ef-4982-ad47-ea206a6faa58\") " pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.251383 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7bccb117-9b78-4a2a-b2fb-34952249e63c-config-volume\") pod \"dns-default-w49nv\" (UID: \"7bccb117-9b78-4a2a-b2fb-34952249e63c\") " pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.252539 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.252871 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/348abc9a-ac32-4c1a-8af5-16a95155591c-srv-cert\") pod \"catalog-operator-68c6474976-zgsc2\" (UID: \"348abc9a-ac32-4c1a-8af5-16a95155591c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.267535 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd"] Oct 07 22:11:10 crc kubenswrapper[4871]: W1007 22:11:10.290493 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5efbd13_6819_4e2c_899f_1711650fbb48.slice/crio-2a970f418a1d4e0af98ea15c1f6e7c01e1a71d20cfda97594f014f8bb1a15a69 WatchSource:0}: Error finding container 2a970f418a1d4e0af98ea15c1f6e7c01e1a71d20cfda97594f014f8bb1a15a69: Status 404 returned error can't find the container with id 2a970f418a1d4e0af98ea15c1f6e7c01e1a71d20cfda97594f014f8bb1a15a69 Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.292377 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctb5q\" (UniqueName: \"kubernetes.io/projected/18a8f350-7724-4302-8663-499480b9af6e-kube-api-access-ctb5q\") pod \"control-plane-machine-set-operator-78cbb6b69f-cc8tr\" (UID: \"18a8f350-7724-4302-8663-499480b9af6e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.301312 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp98p\" (UniqueName: \"kubernetes.io/projected/7bccb117-9b78-4a2a-b2fb-34952249e63c-kube-api-access-lp98p\") pod \"dns-default-w49nv\" (UID: \"7bccb117-9b78-4a2a-b2fb-34952249e63c\") " pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.319350 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srjbm\" (UniqueName: \"kubernetes.io/projected/a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab-kube-api-access-srjbm\") pod \"csi-hostpathplugin-8z4bc\" (UID: \"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.329874 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8nrs6"] Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.332364 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.333505 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvz6c\" (UniqueName: \"kubernetes.io/projected/a265c61b-e26c-40ea-9eee-2fa50b0bbdec-kube-api-access-mvz6c\") pod \"packageserver-d55dfcdfc-wdqtr\" (UID: \"a265c61b-e26c-40ea-9eee-2fa50b0bbdec\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.342030 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.342553 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:10.842529157 +0000 UTC m=+144.645227230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.342646 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.354844 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prp6p\" (UniqueName: \"kubernetes.io/projected/496b361b-01ac-48a6-ab37-c7bea109017d-kube-api-access-prp6p\") pod \"package-server-manager-789f6589d5-4p54k\" (UID: \"496b361b-01ac-48a6-ab37-c7bea109017d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.378111 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cl8z\" (UniqueName: \"kubernetes.io/projected/9d97e32b-02b4-4435-bd25-81174b3e0da9-kube-api-access-8cl8z\") pod \"marketplace-operator-79b997595-kvwx2\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.386461 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.388932 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg"] Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.389408 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.391091 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.401696 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkncv\" (UniqueName: \"kubernetes.io/projected/fd949a6d-e0ef-4982-ad47-ea206a6faa58-kube-api-access-lkncv\") pod \"service-ca-9c57cc56f-x5mfq\" (UID: \"fd949a6d-e0ef-4982-ad47-ea206a6faa58\") " pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.424177 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq5cq\" (UniqueName: \"kubernetes.io/projected/348abc9a-ac32-4c1a-8af5-16a95155591c-kube-api-access-tq5cq\") pod \"catalog-operator-68c6474976-zgsc2\" (UID: \"348abc9a-ac32-4c1a-8af5-16a95155591c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.441966 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw"] Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.443480 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.443946 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:10.943931428 +0000 UTC m=+144.746629501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.452152 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.467650 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.489107 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qctsz\" (UniqueName: \"kubernetes.io/projected/57a0865b-bdb2-4a39-a750-e1b5e77b2e64-kube-api-access-qctsz\") pod \"service-ca-operator-777779d784-cv9kf\" (UID: \"57a0865b-bdb2-4a39-a750-e1b5e77b2e64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.510189 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz2kr\" (UniqueName: \"kubernetes.io/projected/1a916502-6bc9-48ba-906e-831e6fe85e3b-kube-api-access-bz2kr\") pod \"collect-profiles-29331240-6x9cw\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.518893 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rftmz\" (UniqueName: \"kubernetes.io/projected/f6d85de4-a099-42e6-bb04-7016ed806e86-kube-api-access-rftmz\") pod \"machine-config-server-rjlqg\" (UID: \"f6d85de4-a099-42e6-bb04-7016ed806e86\") " pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.524594 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j"] Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.530375 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r"] Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.540901 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd24m\" (UniqueName: \"kubernetes.io/projected/e9ec82f3-8972-4d39-b35d-3f2901beb829-kube-api-access-nd24m\") pod \"olm-operator-6b444d44fb-5722h\" (UID: \"e9ec82f3-8972-4d39-b35d-3f2901beb829\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.542096 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f95pf\" (UniqueName: \"kubernetes.io/projected/30ac8190-e410-439f-9d15-95a62b9096b6-kube-api-access-f95pf\") pod \"ingress-canary-8hxk6\" (UID: \"30ac8190-e410-439f-9d15-95a62b9096b6\") " pod="openshift-ingress-canary/ingress-canary-8hxk6" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.545492 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.545899 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.045874293 +0000 UTC m=+144.848572366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.647721 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.648267 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.148247 +0000 UTC m=+144.950945073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.689743 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.691953 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.697952 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.705077 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.713181 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.754502 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rjlqg" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.755221 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.756475 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.256435922 +0000 UTC m=+145.059133985 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.756522 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8"] Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.762864 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8hxk6" Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.858618 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.859110 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.359097116 +0000 UTC m=+145.161795189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.970146 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.975568 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.475498906 +0000 UTC m=+145.278208920 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:10 crc kubenswrapper[4871]: I1007 22:11:10.975736 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:10 crc kubenswrapper[4871]: E1007 22:11:10.977175 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.477165911 +0000 UTC m=+145.279863984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.043129 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bwsvx"] Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.044991 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf"] Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.077469 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.077758 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.577730059 +0000 UTC m=+145.380428132 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.077856 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.078303 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.578294414 +0000 UTC m=+145.380992487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.115451 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76"] Oct 07 22:11:11 crc kubenswrapper[4871]: W1007 22:11:11.133641 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6afae383_3eb0_4bb9_8da0_6fa74ba1d09a.slice/crio-f642f9be03f7e6cf1c0d592d2e336485b4db0561ab565bfb9e31fc35ef8b075d WatchSource:0}: Error finding container f642f9be03f7e6cf1c0d592d2e336485b4db0561ab565bfb9e31fc35ef8b075d: Status 404 returned error can't find the container with id f642f9be03f7e6cf1c0d592d2e336485b4db0561ab565bfb9e31fc35ef8b075d Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.179067 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.179595 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.679570982 +0000 UTC m=+145.482269065 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.218514 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j" event={"ID":"e38a71c3-e5a2-4460-bc74-fc17bb897f67","Type":"ContainerStarted","Data":"88fc1f87c0b7db6ae05ce9d17895b0a5e12bf6f1fe3af985cafb2fdc0d5999ea"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.261550 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" event={"ID":"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a","Type":"ContainerStarted","Data":"f642f9be03f7e6cf1c0d592d2e336485b4db0561ab565bfb9e31fc35ef8b075d"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.281266 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.282169 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.782156084 +0000 UTC m=+145.584854147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.335301 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" event={"ID":"c085b695-1392-45eb-8fab-2813c719dfac","Type":"ContainerStarted","Data":"e031065b961c0a0c176d5ba2ed04cdd30c58ecd7f0f71fc63f1065cb742fa42c"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.335412 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" event={"ID":"c085b695-1392-45eb-8fab-2813c719dfac","Type":"ContainerStarted","Data":"dc2440fbe606aa2a108f0574cd545e8e3322ae06a18b90ae89c52a409f63b00e"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.342072 4871 generic.go:334] "Generic (PLEG): container finished" podID="f0312d2c-53d5-4916-9956-418acb3ff0b8" containerID="7471e7c38f6fdf1e788e676ac9a3e290a9a034615bc91aacbf13bdcbea5afa15" exitCode=0 Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.342340 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" event={"ID":"f0312d2c-53d5-4916-9956-418acb3ff0b8","Type":"ContainerDied","Data":"7471e7c38f6fdf1e788e676ac9a3e290a9a034615bc91aacbf13bdcbea5afa15"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.368891 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" event={"ID":"05f44398-4018-4ab3-a8c9-93f06d9b85cf","Type":"ContainerStarted","Data":"8b22704c4083a104ee0589bb411ec335536b27a8d6401d5485f22e7de3716e45"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.376545 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" event={"ID":"b4d26eda-8425-4a92-aeb1-418ea15e53fa","Type":"ContainerStarted","Data":"b7d94c397ff471687c5c81870c7ef40764519a5c532babf78840d9a27a39cce9"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.376809 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k"] Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.389627 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.390998 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.890967243 +0000 UTC m=+145.693665316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.451088 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" event={"ID":"c5efbd13-6819-4e2c-899f-1711650fbb48","Type":"ContainerStarted","Data":"2a970f418a1d4e0af98ea15c1f6e7c01e1a71d20cfda97594f014f8bb1a15a69"} Oct 07 22:11:11 crc kubenswrapper[4871]: W1007 22:11:11.460749 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod349b87a8_e51b_4e41_a25d_c8eea29cca90.slice/crio-75e19d9c89f902141cb93b96d3144f38cff978708ca9a5243b51e8e875fbaab4 WatchSource:0}: Error finding container 75e19d9c89f902141cb93b96d3144f38cff978708ca9a5243b51e8e875fbaab4: Status 404 returned error can't find the container with id 75e19d9c89f902141cb93b96d3144f38cff978708ca9a5243b51e8e875fbaab4 Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.466446 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" event={"ID":"252cfed8-c315-49d0-911c-425cfb86fabb","Type":"ContainerStarted","Data":"9f7f8b72885d225ceb41d2be651914e127b54a9bca13121cba2d4e2a92d2fbd7"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.488751 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" event={"ID":"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc","Type":"ContainerStarted","Data":"86bf5d5211e38043fdf2f6e21819d49e4cdd3947789deb35a060c3d43e446f98"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.492875 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.494369 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:11.994348196 +0000 UTC m=+145.797046269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.502841 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-tw5sf"] Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.546458 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" event={"ID":"2a917bbb-dee4-4c7d-852e-12f7be29bf33","Type":"ContainerStarted","Data":"832822dc858734eeacb7a2016f7eb23cd9c49dde96436a0d33d7e2e466a74cdc"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.546526 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" event={"ID":"2a917bbb-dee4-4c7d-852e-12f7be29bf33","Type":"ContainerStarted","Data":"9da70dc42dba1bed8eed1dcfdb1baca67ba4c1c3ccb3adc496d333ca4e5dc80d"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.594235 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.595621 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.095581833 +0000 UTC m=+145.898279896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.618707 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9"] Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.644507 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r8db4" event={"ID":"b1bb9823-6bb1-4e63-8729-efed71f79529","Type":"ContainerStarted","Data":"197963a6e4b40124baa3dd5709aac9ec9eb56deeb0846e81a05a76502be5d0f1"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.683567 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" event={"ID":"c6d38885-8f9a-4840-832d-e831e5027583","Type":"ContainerStarted","Data":"b375c80804723854ee05b8eec6daeb384c742e40bfb66573dbe16a5d0b8cdd96"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.700575 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.702135 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.20211952 +0000 UTC m=+146.004817593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.732486 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" event={"ID":"a6b7f864-334b-4d77-a8e4-bf62b614fb80","Type":"ContainerStarted","Data":"7035d6433c30e5285980a6cacf26e430eb8f122f2c1c565e505cfa09b3a1e65b"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.742029 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" event={"ID":"0609d7b1-7443-4bcd-935e-ca2a61a359b7","Type":"ContainerStarted","Data":"f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.743992 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" event={"ID":"12066056-05f8-4137-8977-f5287bc0d712","Type":"ContainerStarted","Data":"b1e252df215d00dcd1750fb84e1e4166a42bede2208eed530b061b887ff362dc"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.748095 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.751941 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" event={"ID":"554a5462-ff56-4c07-b70b-0ff0f696f988","Type":"ContainerStarted","Data":"89fcdce59be5718924e503f6a6ca5f2d1d680a1f34326246362d705ed3c0344c"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.755923 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" event={"ID":"765ad9da-3b9a-4647-a3ae-b68cee03e99b","Type":"ContainerStarted","Data":"6981a5f8d272b961765f1a56629e7a1f2d5d4a45d65549be23467996c2ddcec5"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.782230 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" event={"ID":"55ee5574-d8f1-48bd-83c3-7aa62e6366f1","Type":"ContainerStarted","Data":"e5e61a947520c54b97237fc7bc682569fa231ce8c9073ae2df83f59dfb7cba6a"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.785017 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.805726 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.805992 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.305963486 +0000 UTC m=+146.108661559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.806177 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.808607 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.308583566 +0000 UTC m=+146.111281839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.838476 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" event={"ID":"f031a182-f40f-4369-83fa-51ba2f2ebd83","Type":"ContainerStarted","Data":"d7d9d409e1bd4247dab57ab86ef337d65c1f6e7250d82dfbbf36d05a384f2500"} Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.839298 4871 patch_prober.go:28] interesting pod/downloads-7954f5f757-hsm2n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.843573 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.847856 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hsm2n" podUID="ff939f3b-6f07-4632-930b-0fa62db4d938" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.854649 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.909907 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.910314 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.410280415 +0000 UTC m=+146.212978478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.910545 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:11 crc kubenswrapper[4871]: E1007 22:11:11.912153 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.412126984 +0000 UTC m=+146.214825057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.959721 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.961607 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 07 22:11:11 crc kubenswrapper[4871]: I1007 22:11:11.961668 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.004844 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-8vdbw" podStartSLOduration=124.004818523 podStartE2EDuration="2m4.004818523s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.002901502 +0000 UTC m=+145.805599575" watchObservedRunningTime="2025-10-07 22:11:12.004818523 +0000 UTC m=+145.807516756" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.012244 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.012356 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.512330313 +0000 UTC m=+146.315028376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.035273 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.083357 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k"] Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.086718 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.586691564 +0000 UTC m=+146.389389637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.105764 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" podStartSLOduration=123.105728471 podStartE2EDuration="2m3.105728471s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.085456291 +0000 UTC m=+145.888154364" watchObservedRunningTime="2025-10-07 22:11:12.105728471 +0000 UTC m=+145.908426544" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.138547 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.139536 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.639511821 +0000 UTC m=+146.442209894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.140946 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr"] Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.141284 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-hsm2n" podStartSLOduration=124.141263797 podStartE2EDuration="2m4.141263797s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.1376174 +0000 UTC m=+145.940315473" watchObservedRunningTime="2025-10-07 22:11:12.141263797 +0000 UTC m=+145.943961870" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.169623 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kvwx2"] Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.227324 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw"] Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.231406 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8z4bc"] Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.240785 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.241163 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-w49nv"] Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.241171 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.741157498 +0000 UTC m=+146.543855571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.258660 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-5gxf8" podStartSLOduration=124.258638214 podStartE2EDuration="2m4.258638214s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.257054351 +0000 UTC m=+146.059752424" watchObservedRunningTime="2025-10-07 22:11:12.258638214 +0000 UTC m=+146.061336287" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.286733 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" podStartSLOduration=124.286710371 podStartE2EDuration="2m4.286710371s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.284885623 +0000 UTC m=+146.087583686" watchObservedRunningTime="2025-10-07 22:11:12.286710371 +0000 UTC m=+146.089408444" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.314361 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.319183 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" podStartSLOduration=124.319171226 podStartE2EDuration="2m4.319171226s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.316491515 +0000 UTC m=+146.119189588" watchObservedRunningTime="2025-10-07 22:11:12.319171226 +0000 UTC m=+146.121869299" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.330408 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf"] Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.342492 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.345070 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.845024405 +0000 UTC m=+146.647722478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.347352 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.347931 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.847913531 +0000 UTC m=+146.650611604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.359095 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" podStartSLOduration=123.359076769 podStartE2EDuration="2m3.359076769s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.344355367 +0000 UTC m=+146.147053440" watchObservedRunningTime="2025-10-07 22:11:12.359076769 +0000 UTC m=+146.161774842" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.370871 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr"] Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.379589 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w74x7" podStartSLOduration=124.379572515 podStartE2EDuration="2m4.379572515s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.377374076 +0000 UTC m=+146.180072149" watchObservedRunningTime="2025-10-07 22:11:12.379572515 +0000 UTC m=+146.182270588" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.417852 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j66lt" podStartSLOduration=124.417831674 podStartE2EDuration="2m4.417831674s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.415372668 +0000 UTC m=+146.218070741" watchObservedRunningTime="2025-10-07 22:11:12.417831674 +0000 UTC m=+146.220529747" Oct 07 22:11:12 crc kubenswrapper[4871]: W1007 22:11:12.447639 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bccb117_9b78_4a2a_b2fb_34952249e63c.slice/crio-e5fd20680da565a33db8fdd2c4d49e8189228f742fb56ef573bf518d1751fb55 WatchSource:0}: Error finding container e5fd20680da565a33db8fdd2c4d49e8189228f742fb56ef573bf518d1751fb55: Status 404 returned error can't find the container with id e5fd20680da565a33db8fdd2c4d49e8189228f742fb56ef573bf518d1751fb55 Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.448234 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.451691 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:12.951660895 +0000 UTC m=+146.754358968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.456323 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-r8db4" podStartSLOduration=123.456296278 podStartE2EDuration="2m3.456296278s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.452163468 +0000 UTC m=+146.254861541" watchObservedRunningTime="2025-10-07 22:11:12.456296278 +0000 UTC m=+146.258994351" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.504078 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-np4z5" podStartSLOduration=124.50404936 podStartE2EDuration="2m4.50404936s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.472922561 +0000 UTC m=+146.275620634" watchObservedRunningTime="2025-10-07 22:11:12.50404936 +0000 UTC m=+146.306747423" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.526641 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" podStartSLOduration=124.526618821 podStartE2EDuration="2m4.526618821s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.506974568 +0000 UTC m=+146.309672641" watchObservedRunningTime="2025-10-07 22:11:12.526618821 +0000 UTC m=+146.329316894" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.539115 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h"] Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.552968 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.553457 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.053443976 +0000 UTC m=+146.856142049 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.558661 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" podStartSLOduration=124.558630414 podStartE2EDuration="2m4.558630414s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.551484554 +0000 UTC m=+146.354182627" watchObservedRunningTime="2025-10-07 22:11:12.558630414 +0000 UTC m=+146.361328487" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.596159 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2"] Oct 07 22:11:12 crc kubenswrapper[4871]: W1007 22:11:12.622757 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9ec82f3_8972_4d39_b35d_3f2901beb829.slice/crio-dea16e2546579faa834a1ce8830e969cc10d1340d1cf9ec63ad399801d4ae6f7 WatchSource:0}: Error finding container dea16e2546579faa834a1ce8830e969cc10d1340d1cf9ec63ad399801d4ae6f7: Status 404 returned error can't find the container with id dea16e2546579faa834a1ce8830e969cc10d1340d1cf9ec63ad399801d4ae6f7 Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.653436 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-2zz24" podStartSLOduration=123.653405878 podStartE2EDuration="2m3.653405878s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:12.643986798 +0000 UTC m=+146.446684881" watchObservedRunningTime="2025-10-07 22:11:12.653405878 +0000 UTC m=+146.456103951" Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.653955 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.654112 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.154091117 +0000 UTC m=+146.956789190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.654182 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.654554 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.154547959 +0000 UTC m=+146.957246032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.675092 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-x5mfq"] Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.690450 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8hxk6"] Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.759211 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.759928 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.259894465 +0000 UTC m=+147.062592538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: W1007 22:11:12.820857 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod348abc9a_ac32_4c1a_8af5_16a95155591c.slice/crio-c94df79a74ccda1c24fd7445b982e10e281ef4c04da9d86050c181ad8f2d5131 WatchSource:0}: Error finding container c94df79a74ccda1c24fd7445b982e10e281ef4c04da9d86050c181ad8f2d5131: Status 404 returned error can't find the container with id c94df79a74ccda1c24fd7445b982e10e281ef4c04da9d86050c181ad8f2d5131 Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.861026 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.861593 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.361579573 +0000 UTC m=+147.164277646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.902283 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" event={"ID":"05f44398-4018-4ab3-a8c9-93f06d9b85cf","Type":"ContainerStarted","Data":"9a7ee8758019dc0e87c198d37192ba0a3532e80f557fd1c4121aff49fc50f9a6"} Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.962269 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:12 crc kubenswrapper[4871]: E1007 22:11:12.963072 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.463056726 +0000 UTC m=+147.265754799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.973884 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r8db4" event={"ID":"b1bb9823-6bb1-4e63-8729-efed71f79529","Type":"ContainerStarted","Data":"6adbb2a9d01ebb13a88608a5f70f848c629c99418d61c33d53443d41ab29927c"} Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.975236 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 22:11:12 crc kubenswrapper[4871]: [-]has-synced failed: reason withheld Oct 07 22:11:12 crc kubenswrapper[4871]: [+]process-running ok Oct 07 22:11:12 crc kubenswrapper[4871]: healthz check failed Oct 07 22:11:12 crc kubenswrapper[4871]: I1007 22:11:12.975297 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.015369 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" event={"ID":"349b87a8-e51b-4e41-a25d-c8eea29cca90","Type":"ContainerStarted","Data":"d588feb67d34242e1d7f079541c1a4d2dbd02a170939a0f678634a1a7dc093b7"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.015452 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" event={"ID":"349b87a8-e51b-4e41-a25d-c8eea29cca90","Type":"ContainerStarted","Data":"75e19d9c89f902141cb93b96d3144f38cff978708ca9a5243b51e8e875fbaab4"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.052923 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8f69k" podStartSLOduration=124.052895709 podStartE2EDuration="2m4.052895709s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.052427887 +0000 UTC m=+146.855125960" watchObservedRunningTime="2025-10-07 22:11:13.052895709 +0000 UTC m=+146.855593782" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.067774 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:13 crc kubenswrapper[4871]: E1007 22:11:13.070333 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.570316403 +0000 UTC m=+147.373014466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.092445 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" event={"ID":"ecf86dbe-b51a-4228-a5ff-959dfe877387","Type":"ContainerStarted","Data":"30d7e5a3d603006320266185fd252d8f81aec6ce44db6e6b27a5def17cb10178"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.092621 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" event={"ID":"ecf86dbe-b51a-4228-a5ff-959dfe877387","Type":"ContainerStarted","Data":"c62e6a844883cbc5c05c5d42162ed1aa47de3b669eef79e1e4421fd2eafb6c7e"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.119210 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" event={"ID":"6afae383-3eb0-4bb9-8da0-6fa74ba1d09a","Type":"ContainerStarted","Data":"f43aa712b47ebd216e51e8890e32020637f9a890fd55b44e68541b02cf29d325"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.133679 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" podStartSLOduration=124.13365896 podStartE2EDuration="2m4.13365896s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.13289192 +0000 UTC m=+146.935589993" watchObservedRunningTime="2025-10-07 22:11:13.13365896 +0000 UTC m=+146.936357033" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.151464 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" event={"ID":"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab","Type":"ContainerStarted","Data":"cb55efac0d399f9274273ed1abf31c0257cb948062d28e92373089b39f7561c0"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.169097 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:13 crc kubenswrapper[4871]: E1007 22:11:13.170242 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.670219944 +0000 UTC m=+147.472918017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.257287 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tw5sf" event={"ID":"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0","Type":"ContainerStarted","Data":"7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.257723 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tw5sf" event={"ID":"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0","Type":"ContainerStarted","Data":"eb24c72fa1fa40f5278749216bd03264ac337c7a54c25f04b190ee69ff201550"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.271923 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:13 crc kubenswrapper[4871]: E1007 22:11:13.272322 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.772305363 +0000 UTC m=+147.575003436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.340263 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" event={"ID":"1a916502-6bc9-48ba-906e-831e6fe85e3b","Type":"ContainerStarted","Data":"d4b10e6afa8ba93e6b6f0991ec6b876659f82252aa9380d4fc4dbe33d728d029"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.342381 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" event={"ID":"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0","Type":"ContainerStarted","Data":"e61bfc0efcff5521c53796254ef0afeeff0d88a92030d7d21d6531f3728fc817"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.342405 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" event={"ID":"159a0fd3-0ddb-4e50-bd97-3d9fc4d9fcb0","Type":"ContainerStarted","Data":"3eec23f1d7c6c785e9f86fda89f68e763782e92847a67f4995b24ec1d8391f69"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.342445 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-tw5sf" podStartSLOduration=125.342423721 podStartE2EDuration="2m5.342423721s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.340612802 +0000 UTC m=+147.143310875" watchObservedRunningTime="2025-10-07 22:11:13.342423721 +0000 UTC m=+147.145121794" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.343011 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-57nxf" podStartSLOduration=124.343003306 podStartE2EDuration="2m4.343003306s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.176557823 +0000 UTC m=+146.979255896" watchObservedRunningTime="2025-10-07 22:11:13.343003306 +0000 UTC m=+147.145701379" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.381507 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:13 crc kubenswrapper[4871]: E1007 22:11:13.383570 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.883546846 +0000 UTC m=+147.686244919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.387555 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" event={"ID":"12066056-05f8-4137-8977-f5287bc0d712","Type":"ContainerStarted","Data":"85984b417093aff5fb83cfde9f6d3c61b497c980541539e95beb0357b0159d80"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.418598 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bwsvx" event={"ID":"7e7841e7-85ae-4d41-acf4-a4603f33d6d4","Type":"ContainerStarted","Data":"59867fa47dd45d8389d9ddf3b5b8cab28d0f949756d1e722ebde72f2cfaa80ac"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.418712 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bwsvx" event={"ID":"7e7841e7-85ae-4d41-acf4-a4603f33d6d4","Type":"ContainerStarted","Data":"6253cf81d6a4a260962372668677d3e7c6513cf0f25db625ce9fe9ce26e3fbaf"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.419614 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.422314 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" event={"ID":"18a8f350-7724-4302-8663-499480b9af6e","Type":"ContainerStarted","Data":"025e83e7a8b0b567c139a6e5a28eede8e32f5c359702804511903e29794a1ea1"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.422891 4871 patch_prober.go:28] interesting pod/console-operator-58897d9998-bwsvx container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.422934 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-bwsvx" podUID="7e7841e7-85ae-4d41-acf4-a4603f33d6d4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.424096 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" event={"ID":"348abc9a-ac32-4c1a-8af5-16a95155591c","Type":"ContainerStarted","Data":"c94df79a74ccda1c24fd7445b982e10e281ef4c04da9d86050c181ad8f2d5131"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.432239 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" event={"ID":"496b361b-01ac-48a6-ab37-c7bea109017d","Type":"ContainerStarted","Data":"f08578b99b91b9b6a32076c2e45a40b661b7fc377e04965867a8eee54ba3ea4c"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.433655 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" event={"ID":"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc","Type":"ContainerStarted","Data":"219c0b2be379c55c4b2c63a5da38635875531cec9939784c1cc985428224de51"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.467090 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" event={"ID":"e9ec82f3-8972-4d39-b35d-3f2901beb829","Type":"ContainerStarted","Data":"dea16e2546579faa834a1ce8830e969cc10d1340d1cf9ec63ad399801d4ae6f7"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.479208 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" event={"ID":"f0312d2c-53d5-4916-9956-418acb3ff0b8","Type":"ContainerStarted","Data":"605575ae106862157bc5ca8f0b0fbf69e4bd69b143555e333a4af357694c2465"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.486137 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:13 crc kubenswrapper[4871]: E1007 22:11:13.491232 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:13.991191773 +0000 UTC m=+147.793889846 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.493638 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8hxk6" event={"ID":"30ac8190-e410-439f-9d15-95a62b9096b6","Type":"ContainerStarted","Data":"19cf8c21f37eb0d13203840e4743b71263cc8812040ff226a194e79cdccb0b43"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.507567 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mvq76" podStartSLOduration=125.507539868 podStartE2EDuration="2m5.507539868s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.398676089 +0000 UTC m=+147.201374152" watchObservedRunningTime="2025-10-07 22:11:13.507539868 +0000 UTC m=+147.310237941" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.542558 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" event={"ID":"c5efbd13-6819-4e2c-899f-1711650fbb48","Type":"ContainerStarted","Data":"7eb1ab917788b6a885fd32ec31eda6e27fca968fcd4c14835826c7069fe904a9"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.542630 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" event={"ID":"c5efbd13-6819-4e2c-899f-1711650fbb48","Type":"ContainerStarted","Data":"7a3ce3c4c0168526b2df8ec759fea735a5185227abbcaeb9830969c121b975a4"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.544306 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-bwsvx" podStartSLOduration=125.544276157 podStartE2EDuration="2m5.544276157s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.501387445 +0000 UTC m=+147.304085508" watchObservedRunningTime="2025-10-07 22:11:13.544276157 +0000 UTC m=+147.346974230" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.580411 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" podStartSLOduration=124.580382679 podStartE2EDuration="2m4.580382679s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.577172653 +0000 UTC m=+147.379870726" watchObservedRunningTime="2025-10-07 22:11:13.580382679 +0000 UTC m=+147.383080752" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.581321 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" event={"ID":"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc","Type":"ContainerStarted","Data":"bf951d0c4edd722ace96da270c26fab259aff6956e153646fac425f8fc685956"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.582405 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" event={"ID":"a0825d5d-ec58-4a42-85e6-7678c5a6a5bc","Type":"ContainerStarted","Data":"19ba5c3195f659a9c8e8a77d19868884c31b09df3d4ffdb71de5cbfae3b0d589"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.588658 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:13 crc kubenswrapper[4871]: E1007 22:11:13.590442 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:14.090414826 +0000 UTC m=+147.893112899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.621455 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rjlqg" event={"ID":"f6d85de4-a099-42e6-bb04-7016ed806e86","Type":"ContainerStarted","Data":"8c91f94b15e7d68bb0acc988225fe50ca03cb2ec9518e043d313446537d246aa"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.621517 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rjlqg" event={"ID":"f6d85de4-a099-42e6-bb04-7016ed806e86","Type":"ContainerStarted","Data":"2d8d13f39123369023641123ff58853fed27a4f2a2a5c06a8937c79f9a5ffeb8"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.655299 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j" event={"ID":"e38a71c3-e5a2-4460-bc74-fc17bb897f67","Type":"ContainerStarted","Data":"55fa3c936aaa0f36a53afcc40569c8f5126c271152e1786ee5a6a458d9639bf4"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.674617 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" podStartSLOduration=124.674595378 podStartE2EDuration="2m4.674595378s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.672552094 +0000 UTC m=+147.475250167" watchObservedRunningTime="2025-10-07 22:11:13.674595378 +0000 UTC m=+147.477293451" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.689690 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.690603 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.691902 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.692822 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" event={"ID":"9d97e32b-02b4-4435-bd25-81174b3e0da9","Type":"ContainerStarted","Data":"43fa43e7834ba9c6349858e1a3d5ca4c6fdfa5e1343b1eea6b2ddec55b325730"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.694579 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:13 crc kubenswrapper[4871]: E1007 22:11:13.694752 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:14.194735244 +0000 UTC m=+147.997433307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.704015 4871 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kvwx2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.704111 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" podUID="9d97e32b-02b4-4435-bd25-81174b3e0da9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.708146 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hkwsg" event={"ID":"765ad9da-3b9a-4647-a3ae-b68cee03e99b","Type":"ContainerStarted","Data":"cca404dd61daa3565f2e7e077a8d71e25a91f0d1215b23cf581741936738b921"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.726523 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" event={"ID":"a265c61b-e26c-40ea-9eee-2fa50b0bbdec","Type":"ContainerStarted","Data":"952eca6eabdd91a435720068c3d9b468acf4f0210909020de6e1d1d8991ef101"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.727948 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.732892 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" event={"ID":"fd949a6d-e0ef-4982-ad47-ea206a6faa58","Type":"ContainerStarted","Data":"439a8cb2bee73ae4b4d90798d10a11e42d38ed2958b0bc95d2fe5212c6b219af"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.735006 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f5tkt" event={"ID":"554a5462-ff56-4c07-b70b-0ff0f696f988","Type":"ContainerStarted","Data":"12361008da28ea133febd99c8094b4a05f5fc3d00990ae8b2acd198f5d04e5d1"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.738832 4871 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wdqtr container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.738883 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" podUID="a265c61b-e26c-40ea-9eee-2fa50b0bbdec" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.761496 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" podStartSLOduration=124.761475121 podStartE2EDuration="2m4.761475121s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.759413396 +0000 UTC m=+147.562111469" watchObservedRunningTime="2025-10-07 22:11:13.761475121 +0000 UTC m=+147.564173194" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.761714 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w49nv" event={"ID":"7bccb117-9b78-4a2a-b2fb-34952249e63c","Type":"ContainerStarted","Data":"e5fd20680da565a33db8fdd2c4d49e8189228f742fb56ef573bf518d1751fb55"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.792778 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" event={"ID":"c6d38885-8f9a-4840-832d-e831e5027583","Type":"ContainerStarted","Data":"cda32b8f49ceeba5ccb7d46292febd1c7167e6891612ec1297cbcd3407c502fc"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.797064 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:13 crc kubenswrapper[4871]: E1007 22:11:13.798297 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:14.298270091 +0000 UTC m=+148.100968164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.826954 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" event={"ID":"57a0865b-bdb2-4a39-a750-e1b5e77b2e64","Type":"ContainerStarted","Data":"aac43bf9c35b6098bc7b8bd4a1a9e88404026e8d369431b622c958c661f8fccb"} Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.856519 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9l8jh" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.858688 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" podStartSLOduration=124.85867348 podStartE2EDuration="2m4.85867348s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.858615568 +0000 UTC m=+147.661313641" watchObservedRunningTime="2025-10-07 22:11:13.85867348 +0000 UTC m=+147.661371553" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.898815 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:13 crc kubenswrapper[4871]: E1007 22:11:13.902118 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:14.402098717 +0000 UTC m=+148.204796980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.963136 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8x2nd" podStartSLOduration=124.963113782 podStartE2EDuration="2m4.963113782s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.961112758 +0000 UTC m=+147.763810831" watchObservedRunningTime="2025-10-07 22:11:13.963113782 +0000 UTC m=+147.765811855" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.964405 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 22:11:13 crc kubenswrapper[4871]: [-]has-synced failed: reason withheld Oct 07 22:11:13 crc kubenswrapper[4871]: [+]process-running ok Oct 07 22:11:13 crc kubenswrapper[4871]: healthz check failed Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.964482 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:13 crc kubenswrapper[4871]: I1007 22:11:13.993175 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2f6w8" podStartSLOduration=124.993145732 podStartE2EDuration="2m4.993145732s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:13.989779642 +0000 UTC m=+147.792477715" watchObservedRunningTime="2025-10-07 22:11:13.993145732 +0000 UTC m=+147.795843815" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.000606 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.001042 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:14.501020841 +0000 UTC m=+148.303718914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.030584 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" podStartSLOduration=125.030553838 podStartE2EDuration="2m5.030553838s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:14.024315522 +0000 UTC m=+147.827013595" watchObservedRunningTime="2025-10-07 22:11:14.030553838 +0000 UTC m=+147.833251911" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.097632 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j" podStartSLOduration=125.097613644 podStartE2EDuration="2m5.097613644s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:14.061986855 +0000 UTC m=+147.864684928" watchObservedRunningTime="2025-10-07 22:11:14.097613644 +0000 UTC m=+147.900311717" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.098873 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-rjlqg" podStartSLOduration=7.098868098 podStartE2EDuration="7.098868098s" podCreationTimestamp="2025-10-07 22:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:14.08732744 +0000 UTC m=+147.890025513" watchObservedRunningTime="2025-10-07 22:11:14.098868098 +0000 UTC m=+147.901566171" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.101910 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.102636 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:14.602287359 +0000 UTC m=+148.404985432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.145462 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w4rlw" podStartSLOduration=125.145443388 podStartE2EDuration="2m5.145443388s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:14.144265817 +0000 UTC m=+147.946963890" watchObservedRunningTime="2025-10-07 22:11:14.145443388 +0000 UTC m=+147.948141461" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.205439 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.205969 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:14.70594887 +0000 UTC m=+148.508646943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.317912 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.318313 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:14.818295862 +0000 UTC m=+148.620993935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.416456 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.416509 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.417921 4871 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-kbkgv container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.418003 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" podUID="f0312d2c-53d5-4916-9956-418acb3ff0b8" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.418676 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.419146 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:14.919125158 +0000 UTC m=+148.721823231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.520332 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.520990 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.02095877 +0000 UTC m=+148.823657064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.621634 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.621839 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.121781746 +0000 UTC m=+148.924479819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.622055 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.622418 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.122403183 +0000 UTC m=+148.925101256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.723427 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.723966 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.223928757 +0000 UTC m=+149.026626820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.825485 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.825959 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.325943954 +0000 UTC m=+149.128642027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.833047 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" event={"ID":"9d97e32b-02b4-4435-bd25-81174b3e0da9","Type":"ContainerStarted","Data":"69605fa35c1363bbb7197afd08ab593a01b606b05174b53b3f359e14c60b2cd3"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.833871 4871 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kvwx2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.834004 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" podUID="9d97e32b-02b4-4435-bd25-81174b3e0da9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.835857 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" event={"ID":"e9ec82f3-8972-4d39-b35d-3f2901beb829","Type":"ContainerStarted","Data":"ad95d6816ecc7b99d04195852d1c29a80683a6d71142673fed1cf8a5512e5cee"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.837173 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.838131 4871 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5722h container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.838169 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" podUID="e9ec82f3-8972-4d39-b35d-3f2901beb829" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.839309 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-m6l6j" event={"ID":"e38a71c3-e5a2-4460-bc74-fc17bb897f67","Type":"ContainerStarted","Data":"76a9ad2b961477c7ec1d5c0b461fdb67396eaa5fafa0ac1a1f5b5d67ff2da260"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.877165 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" podStartSLOduration=125.877145038 podStartE2EDuration="2m5.877145038s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:14.876933652 +0000 UTC m=+148.679631735" watchObservedRunningTime="2025-10-07 22:11:14.877145038 +0000 UTC m=+148.679843101" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.883773 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" event={"ID":"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab","Type":"ContainerStarted","Data":"03af89b0dadebbe6d268d7524573f4b7d3f7828c1f1e553f73b70ceb0e0794cf"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.889742 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" event={"ID":"496b361b-01ac-48a6-ab37-c7bea109017d","Type":"ContainerStarted","Data":"b31778e839f65ff31ded499e16e2f14d7ac92891bc2302c4feae4bf8c4d2cab4"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.889833 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" event={"ID":"496b361b-01ac-48a6-ab37-c7bea109017d","Type":"ContainerStarted","Data":"af09882ff48b1d2c23ee36473106a903872f332dbed657b31f3520fafd7c59af"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.890014 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.893636 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8hxk6" event={"ID":"30ac8190-e410-439f-9d15-95a62b9096b6","Type":"ContainerStarted","Data":"6861efbb140869f6ff1d1a7370bb7a19d8a87492d1785afacf6183eb49d22023"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.908300 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cc8tr" event={"ID":"18a8f350-7724-4302-8663-499480b9af6e","Type":"ContainerStarted","Data":"0de26ca974c8ca4429bbda9621f505eab5d0b6ba5a5fefbf427bd7ed1f1cb5b6"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.919116 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" event={"ID":"348abc9a-ac32-4c1a-8af5-16a95155591c","Type":"ContainerStarted","Data":"a7ace48e67398dda07770a6b45b95ba798e9b2f6ec622dc9f30e52da2ef607d3"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.920040 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.922958 4871 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-zgsc2 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.923031 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" podUID="348abc9a-ac32-4c1a-8af5-16a95155591c" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.926641 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.929159 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" event={"ID":"57a0865b-bdb2-4a39-a750-e1b5e77b2e64","Type":"ContainerStarted","Data":"47adff59c3f99458dce3ddbd74b5d632b03ec4655c152d04599e4330995ccb35"} Oct 07 22:11:14 crc kubenswrapper[4871]: E1007 22:11:14.929388 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.429362039 +0000 UTC m=+149.232060112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.942063 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" event={"ID":"1a916502-6bc9-48ba-906e-831e6fe85e3b","Type":"ContainerStarted","Data":"7a551e5fadb1293eab5adb4685604c781f15988d0b22ba8462832dc5b987e10d"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.964456 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w49nv" event={"ID":"7bccb117-9b78-4a2a-b2fb-34952249e63c","Type":"ContainerStarted","Data":"0479899c8e55e3da13e8057497c87823f528f8aa3a14109f59f02ab482783153"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.964515 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w49nv" event={"ID":"7bccb117-9b78-4a2a-b2fb-34952249e63c","Type":"ContainerStarted","Data":"e4cced7e082b8491a89570d9aa0f4d95bf85f60fe0228486cc6531cdd731e6cc"} Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.965242 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.975487 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 22:11:14 crc kubenswrapper[4871]: [-]has-synced failed: reason withheld Oct 07 22:11:14 crc kubenswrapper[4871]: [+]process-running ok Oct 07 22:11:14 crc kubenswrapper[4871]: healthz check failed Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.975554 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:14 crc kubenswrapper[4871]: I1007 22:11:14.996092 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kqcm9" event={"ID":"341ee697-f8f5-4ee9-b5a2-7e2a8d8426cc","Type":"ContainerStarted","Data":"259026c8c11d9722e781bbb8964dae8e9152f6b08b7608181ed46b66e2e53cb1"} Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.011019 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" event={"ID":"12066056-05f8-4137-8977-f5287bc0d712","Type":"ContainerStarted","Data":"5dbcc4c41ce28208d201d3d2c9d17ccab6d396a48e6bd5967188600f19e1a194"} Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.032347 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.034374 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" podStartSLOduration=126.034359855 podStartE2EDuration="2m6.034359855s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:14.983286075 +0000 UTC m=+148.785984138" watchObservedRunningTime="2025-10-07 22:11:15.034359855 +0000 UTC m=+148.837057928" Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.036583 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.536561554 +0000 UTC m=+149.339259627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.036985 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" event={"ID":"a265c61b-e26c-40ea-9eee-2fa50b0bbdec","Type":"ContainerStarted","Data":"331b11a49089f1be8fa854939ef5f9f995c2f0d1d11ef119883c7d57690b6636"} Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.070879 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" event={"ID":"05f44398-4018-4ab3-a8c9-93f06d9b85cf","Type":"ContainerStarted","Data":"cb8afc49c5a12c556f643fc212c721d30d0c29f2352d30a631feb59ae0a49f66"} Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.096907 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" podStartSLOduration=126.096887611 podStartE2EDuration="2m6.096887611s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:15.094748494 +0000 UTC m=+148.897446567" watchObservedRunningTime="2025-10-07 22:11:15.096887611 +0000 UTC m=+148.899585684" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.104516 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" event={"ID":"fd949a6d-e0ef-4982-ad47-ea206a6faa58","Type":"ContainerStarted","Data":"fb1e20fa8faadafb042534d78499598d5e38017073134016642213953eebe032"} Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.106725 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cv9kf" podStartSLOduration=126.106706082 podStartE2EDuration="2m6.106706082s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:15.035352022 +0000 UTC m=+148.838050095" watchObservedRunningTime="2025-10-07 22:11:15.106706082 +0000 UTC m=+148.909404155" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.133705 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.136770 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.636742632 +0000 UTC m=+149.439440705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.204002 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-w49nv" podStartSLOduration=8.203977863 podStartE2EDuration="8.203977863s" podCreationTimestamp="2025-10-07 22:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:15.203168932 +0000 UTC m=+149.005867005" watchObservedRunningTime="2025-10-07 22:11:15.203977863 +0000 UTC m=+149.006675936" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.236075 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.237570 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.737546807 +0000 UTC m=+149.540244880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.257194 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" podStartSLOduration=126.25716704 podStartE2EDuration="2m6.25716704s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:15.249920077 +0000 UTC m=+149.052618160" watchObservedRunningTime="2025-10-07 22:11:15.25716704 +0000 UTC m=+149.059865113" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.338775 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.338999 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.838955068 +0000 UTC m=+149.641653141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.339479 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.339921 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.839912594 +0000 UTC m=+149.642610667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.340760 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-8hxk6" podStartSLOduration=8.340737616 podStartE2EDuration="8.340737616s" podCreationTimestamp="2025-10-07 22:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:15.292713307 +0000 UTC m=+149.095411380" watchObservedRunningTime="2025-10-07 22:11:15.340737616 +0000 UTC m=+149.143435689" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.441208 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.441460 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.941426088 +0000 UTC m=+149.744124161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.441600 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.442014 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:15.942004443 +0000 UTC m=+149.744702516 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.448264 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-x5mfq" podStartSLOduration=126.44824834 podStartE2EDuration="2m6.44824834s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:15.34089839 +0000 UTC m=+149.143596463" watchObservedRunningTime="2025-10-07 22:11:15.44824834 +0000 UTC m=+149.250946413" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.506031 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-8nrs6" podStartSLOduration=126.505989638 podStartE2EDuration="2m6.505989638s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:15.501920909 +0000 UTC m=+149.304618982" watchObservedRunningTime="2025-10-07 22:11:15.505989638 +0000 UTC m=+149.308687711" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.515720 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cq6r" podStartSLOduration=126.515686396 podStartE2EDuration="2m6.515686396s" podCreationTimestamp="2025-10-07 22:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:15.455182484 +0000 UTC m=+149.257880557" watchObservedRunningTime="2025-10-07 22:11:15.515686396 +0000 UTC m=+149.318384469" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.543357 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.543743 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.043714422 +0000 UTC m=+149.846412495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.647505 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.647945 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.147930328 +0000 UTC m=+149.950628401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.754683 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.755267 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.755299 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.755334 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.755376 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.757106 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.257065915 +0000 UTC m=+150.059763988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.762435 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.762615 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wdqtr" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.786866 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.787049 4871 patch_prober.go:28] interesting pod/apiserver-76f77b778f-5x65s container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]log ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]etcd ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/generic-apiserver-start-informers ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/max-in-flight-filter ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 07 22:11:15 crc kubenswrapper[4871]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 07 22:11:15 crc kubenswrapper[4871]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/project.openshift.io-projectcache ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/openshift.io-startinformers ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 07 22:11:15 crc kubenswrapper[4871]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 07 22:11:15 crc kubenswrapper[4871]: livez check failed Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.787098 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" podUID="ecf86dbe-b51a-4228-a5ff-959dfe877387" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.793357 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.803306 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.815133 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.824116 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.836193 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.859595 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.860052 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.360036428 +0000 UTC m=+150.162734511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.960783 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:15 crc kubenswrapper[4871]: E1007 22:11:15.961162 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.461141961 +0000 UTC m=+150.263840034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.973565 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 22:11:15 crc kubenswrapper[4871]: [-]has-synced failed: reason withheld Oct 07 22:11:15 crc kubenswrapper[4871]: [+]process-running ok Oct 07 22:11:15 crc kubenswrapper[4871]: healthz check failed Oct 07 22:11:15 crc kubenswrapper[4871]: I1007 22:11:15.973647 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.066599 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.067339 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.567323659 +0000 UTC m=+150.370021732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.110877 4871 patch_prober.go:28] interesting pod/console-operator-58897d9998-bwsvx container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.110964 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-bwsvx" podUID="7e7841e7-85ae-4d41-acf4-a4603f33d6d4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.168151 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.168503 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.668482464 +0000 UTC m=+150.471180537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.169379 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" event={"ID":"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab","Type":"ContainerStarted","Data":"ace42a6423570f44f3376f1297b5050781a303e41d1e619fff15cd55425e9456"} Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.176120 4871 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kvwx2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.176174 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" podUID="9d97e32b-02b4-4435-bd25-81174b3e0da9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.194694 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5722h" Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.196614 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-zgsc2" Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.269766 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.274012 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.773989514 +0000 UTC m=+150.576687587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: W1007 22:11:16.351615 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-f0c37fd4d1b291bccedf7794a9f7e1fdb6aebac435ea5bf953f5fb00568b20e2 WatchSource:0}: Error finding container f0c37fd4d1b291bccedf7794a9f7e1fdb6aebac435ea5bf953f5fb00568b20e2: Status 404 returned error can't find the container with id f0c37fd4d1b291bccedf7794a9f7e1fdb6aebac435ea5bf953f5fb00568b20e2 Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.376542 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.376874 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.876853444 +0000 UTC m=+150.679551517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.481304 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.481719 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:16.981704476 +0000 UTC m=+150.784402549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.579753 4871 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.582685 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.582779 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:17.082752968 +0000 UTC m=+150.885451041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.584078 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.584640 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:17.084625288 +0000 UTC m=+150.887323351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.688509 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.688985 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:17.188958937 +0000 UTC m=+150.991657010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.789727 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.790152 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:17.290136882 +0000 UTC m=+151.092834955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.891016 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.891362 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:17.391321187 +0000 UTC m=+151.194019260 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.891451 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.891915 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:17.391906633 +0000 UTC m=+151.194604696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.963921 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 22:11:16 crc kubenswrapper[4871]: [-]has-synced failed: reason withheld Oct 07 22:11:16 crc kubenswrapper[4871]: [+]process-running ok Oct 07 22:11:16 crc kubenswrapper[4871]: healthz check failed Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.964002 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:16 crc kubenswrapper[4871]: I1007 22:11:16.992600 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:16 crc kubenswrapper[4871]: E1007 22:11:16.992955 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:17.492929133 +0000 UTC m=+151.295627196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.116856 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:17 crc kubenswrapper[4871]: E1007 22:11:17.117785 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 22:11:17.617763948 +0000 UTC m=+151.420462021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rfkm7" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.147511 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.186709 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" event={"ID":"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab","Type":"ContainerStarted","Data":"0c5cfce6f36a06b3cd76cfdd4d62dc0a4695be393de75cd364f44512f814dad4"} Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.187056 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" event={"ID":"a57fc2ee-69ec-4e6b-9ed1-70917f8c51ab","Type":"ContainerStarted","Data":"911ecbc215265399423a79cafef1b9083f3a9ea78592b5018cabcba2284c72a1"} Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.192561 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d2490845154a90e3b520611ac78cdf59d98afc0645b58bdee1a06246fe893ac9"} Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.192686 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"91cefd5ad6a5e8afc04b419079b4472bdfe437740ff1717584201de70c837e91"} Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.196075 4871 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-07T22:11:16.583717204Z","Handler":null,"Name":""} Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.199827 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f637f60fd56cf37f857831a4f912cc4605679f112b4809b886119886487326eb"} Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.199902 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f0c37fd4d1b291bccedf7794a9f7e1fdb6aebac435ea5bf953f5fb00568b20e2"} Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.210496 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c4d17b1ce7dc32fcdeff43d5c75c39be6dd23c2636594220ad696232094626df"} Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.210554 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f63cd3b7441af76a58d711a5a8a405d0e6ab20a3693537b136293e8334fe974c"} Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.216034 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-8z4bc" podStartSLOduration=10.216020925 podStartE2EDuration="10.216020925s" podCreationTimestamp="2025-10-07 22:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:17.213087467 +0000 UTC m=+151.015785540" watchObservedRunningTime="2025-10-07 22:11:17.216020925 +0000 UTC m=+151.018718998" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.217730 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:17 crc kubenswrapper[4871]: E1007 22:11:17.218093 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 22:11:17.7180794 +0000 UTC m=+151.520777473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.231296 4871 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.231349 4871 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.319710 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.323048 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.323083 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.367482 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rfkm7\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.421863 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.430158 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.468434 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rwkd7"] Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.469487 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.473442 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.494882 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rwkd7"] Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.499355 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.624824 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-utilities\") pod \"certified-operators-rwkd7\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.625222 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5j22\" (UniqueName: \"kubernetes.io/projected/d02afe1f-19f0-448f-a9a1-0e7f19b39949-kube-api-access-x5j22\") pod \"certified-operators-rwkd7\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.625322 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-catalog-content\") pod \"certified-operators-rwkd7\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.671940 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dqpjz"] Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.673625 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.678601 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dqpjz"] Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.680064 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.723859 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rfkm7"] Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.726896 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-catalog-content\") pod \"certified-operators-rwkd7\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.726973 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-utilities\") pod \"certified-operators-rwkd7\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.727000 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5j22\" (UniqueName: \"kubernetes.io/projected/d02afe1f-19f0-448f-a9a1-0e7f19b39949-kube-api-access-x5j22\") pod \"certified-operators-rwkd7\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.728184 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-catalog-content\") pod \"certified-operators-rwkd7\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.728305 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-utilities\") pod \"certified-operators-rwkd7\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.746270 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5j22\" (UniqueName: \"kubernetes.io/projected/d02afe1f-19f0-448f-a9a1-0e7f19b39949-kube-api-access-x5j22\") pod \"certified-operators-rwkd7\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.784339 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.828099 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-catalog-content\") pod \"community-operators-dqpjz\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.828152 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-utilities\") pod \"community-operators-dqpjz\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.828205 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvqcw\" (UniqueName: \"kubernetes.io/projected/d868b268-45e4-480f-b84c-b74b51127ae6-kube-api-access-qvqcw\") pod \"community-operators-dqpjz\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.869314 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hn2s5"] Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.870268 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.888725 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hn2s5"] Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.930835 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvqcw\" (UniqueName: \"kubernetes.io/projected/d868b268-45e4-480f-b84c-b74b51127ae6-kube-api-access-qvqcw\") pod \"community-operators-dqpjz\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.930965 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-catalog-content\") pod \"community-operators-dqpjz\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.930999 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-utilities\") pod \"community-operators-dqpjz\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.931572 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-utilities\") pod \"community-operators-dqpjz\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.932052 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-catalog-content\") pod \"community-operators-dqpjz\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.969004 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvqcw\" (UniqueName: \"kubernetes.io/projected/d868b268-45e4-480f-b84c-b74b51127ae6-kube-api-access-qvqcw\") pod \"community-operators-dqpjz\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.973079 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 22:11:17 crc kubenswrapper[4871]: [-]has-synced failed: reason withheld Oct 07 22:11:17 crc kubenswrapper[4871]: [+]process-running ok Oct 07 22:11:17 crc kubenswrapper[4871]: healthz check failed Oct 07 22:11:17 crc kubenswrapper[4871]: I1007 22:11:17.973193 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.004022 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.032010 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p94kn\" (UniqueName: \"kubernetes.io/projected/483620a6-7917-403c-af03-8567c258e600-kube-api-access-p94kn\") pod \"certified-operators-hn2s5\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.032128 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-utilities\") pod \"certified-operators-hn2s5\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.032154 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-catalog-content\") pod \"certified-operators-hn2s5\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.063504 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rwkd7"] Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.073727 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dt75j"] Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.074888 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.083133 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dt75j"] Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.133705 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-utilities\") pod \"certified-operators-hn2s5\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.134597 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-catalog-content\") pod \"certified-operators-hn2s5\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.134639 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p94kn\" (UniqueName: \"kubernetes.io/projected/483620a6-7917-403c-af03-8567c258e600-kube-api-access-p94kn\") pod \"certified-operators-hn2s5\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.134555 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-utilities\") pod \"certified-operators-hn2s5\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.135332 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-catalog-content\") pod \"certified-operators-hn2s5\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.163128 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p94kn\" (UniqueName: \"kubernetes.io/projected/483620a6-7917-403c-af03-8567c258e600-kube-api-access-p94kn\") pod \"certified-operators-hn2s5\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.193457 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.242261 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" event={"ID":"3689423d-c1f1-46c8-9ea5-165147a8a286","Type":"ContainerStarted","Data":"51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358"} Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.242329 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" event={"ID":"3689423d-c1f1-46c8-9ea5-165147a8a286","Type":"ContainerStarted","Data":"5b729518d75ab356a8cbda851f0ba071062df62fd0d10ef18e236e225ab99161"} Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.243425 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.249506 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-utilities\") pod \"community-operators-dt75j\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.249835 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-catalog-content\") pod \"community-operators-dt75j\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.249957 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzqqc\" (UniqueName: \"kubernetes.io/projected/813d2075-6f88-4e62-9d27-575353b29476-kube-api-access-gzqqc\") pod \"community-operators-dt75j\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.252347 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwkd7" event={"ID":"d02afe1f-19f0-448f-a9a1-0e7f19b39949","Type":"ContainerStarted","Data":"86f45828fcdfdc87168bf8dd3968cad59215b19e2aa9d96d32dea167aa7902dd"} Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.267137 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" podStartSLOduration=130.267113351 podStartE2EDuration="2m10.267113351s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:18.265550789 +0000 UTC m=+152.068248872" watchObservedRunningTime="2025-10-07 22:11:18.267113351 +0000 UTC m=+152.069811424" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.341073 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dqpjz"] Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.351491 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-utilities\") pod \"community-operators-dt75j\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.351573 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-catalog-content\") pod \"community-operators-dt75j\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.351640 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzqqc\" (UniqueName: \"kubernetes.io/projected/813d2075-6f88-4e62-9d27-575353b29476-kube-api-access-gzqqc\") pod \"community-operators-dt75j\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.352902 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-utilities\") pod \"community-operators-dt75j\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.354678 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-catalog-content\") pod \"community-operators-dt75j\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.371963 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzqqc\" (UniqueName: \"kubernetes.io/projected/813d2075-6f88-4e62-9d27-575353b29476-kube-api-access-gzqqc\") pod \"community-operators-dt75j\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: W1007 22:11:18.400409 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd868b268_45e4_480f_b84c_b74b51127ae6.slice/crio-6f0aa10e85629533fc74aca60bc8b98e1552283c0fe10fc10c5fe66de85ec3ab WatchSource:0}: Error finding container 6f0aa10e85629533fc74aca60bc8b98e1552283c0fe10fc10c5fe66de85ec3ab: Status 404 returned error can't find the container with id 6f0aa10e85629533fc74aca60bc8b98e1552283c0fe10fc10c5fe66de85ec3ab Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.407034 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.491836 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hn2s5"] Oct 07 22:11:18 crc kubenswrapper[4871]: W1007 22:11:18.541781 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod483620a6_7917_403c_af03_8567c258e600.slice/crio-d9a1f827a3f12a1bc704e5b9853d2f445a3e237453c83a724b34dd8bf6850791 WatchSource:0}: Error finding container d9a1f827a3f12a1bc704e5b9853d2f445a3e237453c83a724b34dd8bf6850791: Status 404 returned error can't find the container with id d9a1f827a3f12a1bc704e5b9853d2f445a3e237453c83a724b34dd8bf6850791 Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.685287 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dt75j"] Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.694997 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.700115 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-5x65s" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.802200 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.802982 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.808313 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.808326 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.818366 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.963317 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 22:11:18 crc kubenswrapper[4871]: [-]has-synced failed: reason withheld Oct 07 22:11:18 crc kubenswrapper[4871]: [+]process-running ok Oct 07 22:11:18 crc kubenswrapper[4871]: healthz check failed Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.963707 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.964457 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34f458c3-4674-45bf-a2e6-34d35adbde8c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"34f458c3-4674-45bf-a2e6-34d35adbde8c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:18 crc kubenswrapper[4871]: I1007 22:11:18.964529 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34f458c3-4674-45bf-a2e6-34d35adbde8c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"34f458c3-4674-45bf-a2e6-34d35adbde8c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.002962 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.017685 4871 patch_prober.go:28] interesting pod/downloads-7954f5f757-hsm2n container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.018025 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hsm2n" podUID="ff939f3b-6f07-4632-930b-0fa62db4d938" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.017852 4871 patch_prober.go:28] interesting pod/downloads-7954f5f757-hsm2n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.018370 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hsm2n" podUID="ff939f3b-6f07-4632-930b-0fa62db4d938" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.065554 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34f458c3-4674-45bf-a2e6-34d35adbde8c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"34f458c3-4674-45bf-a2e6-34d35adbde8c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.065612 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34f458c3-4674-45bf-a2e6-34d35adbde8c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"34f458c3-4674-45bf-a2e6-34d35adbde8c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.066052 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34f458c3-4674-45bf-a2e6-34d35adbde8c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"34f458c3-4674-45bf-a2e6-34d35adbde8c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.094097 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34f458c3-4674-45bf-a2e6-34d35adbde8c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"34f458c3-4674-45bf-a2e6-34d35adbde8c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.132048 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.258829 4871 generic.go:334] "Generic (PLEG): container finished" podID="813d2075-6f88-4e62-9d27-575353b29476" containerID="f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa" exitCode=0 Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.258899 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dt75j" event={"ID":"813d2075-6f88-4e62-9d27-575353b29476","Type":"ContainerDied","Data":"f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa"} Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.258948 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dt75j" event={"ID":"813d2075-6f88-4e62-9d27-575353b29476","Type":"ContainerStarted","Data":"013e60ae5f426de2054ba1be4cb236498ad533aaa7607b24243bc7409142ed5c"} Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.260145 4871 generic.go:334] "Generic (PLEG): container finished" podID="1a916502-6bc9-48ba-906e-831e6fe85e3b" containerID="7a551e5fadb1293eab5adb4685604c781f15988d0b22ba8462832dc5b987e10d" exitCode=0 Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.260199 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" event={"ID":"1a916502-6bc9-48ba-906e-831e6fe85e3b","Type":"ContainerDied","Data":"7a551e5fadb1293eab5adb4685604c781f15988d0b22ba8462832dc5b987e10d"} Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.261602 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.262683 4871 generic.go:334] "Generic (PLEG): container finished" podID="d868b268-45e4-480f-b84c-b74b51127ae6" containerID="11921cd710209bea5354890b3de2e53cc9330fd01a36f84d1b0377d5c5a64b81" exitCode=0 Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.262746 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqpjz" event={"ID":"d868b268-45e4-480f-b84c-b74b51127ae6","Type":"ContainerDied","Data":"11921cd710209bea5354890b3de2e53cc9330fd01a36f84d1b0377d5c5a64b81"} Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.262767 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqpjz" event={"ID":"d868b268-45e4-480f-b84c-b74b51127ae6","Type":"ContainerStarted","Data":"6f0aa10e85629533fc74aca60bc8b98e1552283c0fe10fc10c5fe66de85ec3ab"} Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.265777 4871 generic.go:334] "Generic (PLEG): container finished" podID="483620a6-7917-403c-af03-8567c258e600" containerID="c8d18ef5598fc8b50dda841bef89d5bfaeab9102bc3274346f6f6a103efdc365" exitCode=0 Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.265865 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hn2s5" event={"ID":"483620a6-7917-403c-af03-8567c258e600","Type":"ContainerDied","Data":"c8d18ef5598fc8b50dda841bef89d5bfaeab9102bc3274346f6f6a103efdc365"} Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.265898 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hn2s5" event={"ID":"483620a6-7917-403c-af03-8567c258e600","Type":"ContainerStarted","Data":"d9a1f827a3f12a1bc704e5b9853d2f445a3e237453c83a724b34dd8bf6850791"} Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.269363 4871 generic.go:334] "Generic (PLEG): container finished" podID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerID="b52292ec72c1b7b8bded703a8c25d859c5832accdf791a978969b78be1cac7fb" exitCode=0 Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.270132 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwkd7" event={"ID":"d02afe1f-19f0-448f-a9a1-0e7f19b39949","Type":"ContainerDied","Data":"b52292ec72c1b7b8bded703a8c25d859c5832accdf791a978969b78be1cac7fb"} Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.400161 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 22:11:19 crc kubenswrapper[4871]: W1007 22:11:19.408770 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod34f458c3_4674_45bf_a2e6_34d35adbde8c.slice/crio-de7e95af65521655c39046678555e85b06e97ac6da75fc9b5fda3a786e3df0c0 WatchSource:0}: Error finding container de7e95af65521655c39046678555e85b06e97ac6da75fc9b5fda3a786e3df0c0: Status 404 returned error can't find the container with id de7e95af65521655c39046678555e85b06e97ac6da75fc9b5fda3a786e3df0c0 Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.423095 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.429211 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbkgv" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.668559 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pbjpj"] Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.669669 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.672028 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.695251 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbjpj"] Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.776369 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-utilities\") pod \"redhat-marketplace-pbjpj\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.776464 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-catalog-content\") pod \"redhat-marketplace-pbjpj\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.776515 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp6kt\" (UniqueName: \"kubernetes.io/projected/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-kube-api-access-dp6kt\") pod \"redhat-marketplace-pbjpj\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.878096 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-utilities\") pod \"redhat-marketplace-pbjpj\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.878186 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-catalog-content\") pod \"redhat-marketplace-pbjpj\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.878248 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp6kt\" (UniqueName: \"kubernetes.io/projected/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-kube-api-access-dp6kt\") pod \"redhat-marketplace-pbjpj\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.879689 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-catalog-content\") pod \"redhat-marketplace-pbjpj\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.879915 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-utilities\") pod \"redhat-marketplace-pbjpj\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.917863 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp6kt\" (UniqueName: \"kubernetes.io/projected/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-kube-api-access-dp6kt\") pod \"redhat-marketplace-pbjpj\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.959742 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.964354 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 22:11:19 crc kubenswrapper[4871]: [-]has-synced failed: reason withheld Oct 07 22:11:19 crc kubenswrapper[4871]: [+]process-running ok Oct 07 22:11:19 crc kubenswrapper[4871]: healthz check failed Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.964420 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:19 crc kubenswrapper[4871]: I1007 22:11:19.983073 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.053298 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-bwsvx" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.076237 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x95hj"] Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.079589 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.102292 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95hj"] Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.184206 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-utilities\") pod \"redhat-marketplace-x95hj\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.184524 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdx48\" (UniqueName: \"kubernetes.io/projected/f43a1d45-76b6-4fab-aaf4-150c13787a3d-kube-api-access-pdx48\") pod \"redhat-marketplace-x95hj\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.184770 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-catalog-content\") pod \"redhat-marketplace-x95hj\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.247866 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.247931 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.250039 4871 patch_prober.go:28] interesting pod/console-f9d7485db-tw5sf container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.250092 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-tw5sf" podUID="cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.271016 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbjpj"] Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.286531 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-catalog-content\") pod \"redhat-marketplace-x95hj\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.286600 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-utilities\") pod \"redhat-marketplace-x95hj\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.286627 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdx48\" (UniqueName: \"kubernetes.io/projected/f43a1d45-76b6-4fab-aaf4-150c13787a3d-kube-api-access-pdx48\") pod \"redhat-marketplace-x95hj\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.287167 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-utilities\") pod \"redhat-marketplace-x95hj\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.287171 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-catalog-content\") pod \"redhat-marketplace-x95hj\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.314665 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdx48\" (UniqueName: \"kubernetes.io/projected/f43a1d45-76b6-4fab-aaf4-150c13787a3d-kube-api-access-pdx48\") pod \"redhat-marketplace-x95hj\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.329671 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"34f458c3-4674-45bf-a2e6-34d35adbde8c","Type":"ContainerStarted","Data":"d830519f81b99e97eda54435bba44f2aff636773ac05de6dab11019461179590"} Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.329741 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"34f458c3-4674-45bf-a2e6-34d35adbde8c","Type":"ContainerStarted","Data":"de7e95af65521655c39046678555e85b06e97ac6da75fc9b5fda3a786e3df0c0"} Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.349686 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.3496618 podStartE2EDuration="2.3496618s" podCreationTimestamp="2025-10-07 22:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:20.348206582 +0000 UTC m=+154.150904655" watchObservedRunningTime="2025-10-07 22:11:20.3496618 +0000 UTC m=+154.152359873" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.406885 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.415833 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.678570 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.691106 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gjghv"] Oct 07 22:11:20 crc kubenswrapper[4871]: E1007 22:11:20.691501 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a916502-6bc9-48ba-906e-831e6fe85e3b" containerName="collect-profiles" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.691518 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a916502-6bc9-48ba-906e-831e6fe85e3b" containerName="collect-profiles" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.691655 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a916502-6bc9-48ba-906e-831e6fe85e3b" containerName="collect-profiles" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.692576 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.696465 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.706015 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gjghv"] Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.800604 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a916502-6bc9-48ba-906e-831e6fe85e3b-secret-volume\") pod \"1a916502-6bc9-48ba-906e-831e6fe85e3b\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.801596 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a916502-6bc9-48ba-906e-831e6fe85e3b-config-volume\") pod \"1a916502-6bc9-48ba-906e-831e6fe85e3b\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.801632 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bz2kr\" (UniqueName: \"kubernetes.io/projected/1a916502-6bc9-48ba-906e-831e6fe85e3b-kube-api-access-bz2kr\") pod \"1a916502-6bc9-48ba-906e-831e6fe85e3b\" (UID: \"1a916502-6bc9-48ba-906e-831e6fe85e3b\") " Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.802451 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9szcc\" (UniqueName: \"kubernetes.io/projected/91c74861-3e58-4603-b0fd-d791b0368823-kube-api-access-9szcc\") pod \"redhat-operators-gjghv\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.802528 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-catalog-content\") pod \"redhat-operators-gjghv\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.802561 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-utilities\") pod \"redhat-operators-gjghv\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.803544 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95hj"] Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.803616 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a916502-6bc9-48ba-906e-831e6fe85e3b-config-volume" (OuterVolumeSpecName: "config-volume") pod "1a916502-6bc9-48ba-906e-831e6fe85e3b" (UID: "1a916502-6bc9-48ba-906e-831e6fe85e3b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.809064 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a916502-6bc9-48ba-906e-831e6fe85e3b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1a916502-6bc9-48ba-906e-831e6fe85e3b" (UID: "1a916502-6bc9-48ba-906e-831e6fe85e3b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.809314 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a916502-6bc9-48ba-906e-831e6fe85e3b-kube-api-access-bz2kr" (OuterVolumeSpecName: "kube-api-access-bz2kr") pod "1a916502-6bc9-48ba-906e-831e6fe85e3b" (UID: "1a916502-6bc9-48ba-906e-831e6fe85e3b"). InnerVolumeSpecName "kube-api-access-bz2kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.904590 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9szcc\" (UniqueName: \"kubernetes.io/projected/91c74861-3e58-4603-b0fd-d791b0368823-kube-api-access-9szcc\") pod \"redhat-operators-gjghv\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.904695 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-catalog-content\") pod \"redhat-operators-gjghv\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.904724 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-utilities\") pod \"redhat-operators-gjghv\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.904807 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bz2kr\" (UniqueName: \"kubernetes.io/projected/1a916502-6bc9-48ba-906e-831e6fe85e3b-kube-api-access-bz2kr\") on node \"crc\" DevicePath \"\"" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.904820 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a916502-6bc9-48ba-906e-831e6fe85e3b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.904829 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a916502-6bc9-48ba-906e-831e6fe85e3b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.905319 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-utilities\") pod \"redhat-operators-gjghv\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.905425 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-catalog-content\") pod \"redhat-operators-gjghv\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.924407 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9szcc\" (UniqueName: \"kubernetes.io/projected/91c74861-3e58-4603-b0fd-d791b0368823-kube-api-access-9szcc\") pod \"redhat-operators-gjghv\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.963760 4871 patch_prober.go:28] interesting pod/router-default-5444994796-r8db4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 22:11:20 crc kubenswrapper[4871]: [+]has-synced ok Oct 07 22:11:20 crc kubenswrapper[4871]: [+]process-running ok Oct 07 22:11:20 crc kubenswrapper[4871]: healthz check failed Oct 07 22:11:20 crc kubenswrapper[4871]: I1007 22:11:20.964207 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8db4" podUID="b1bb9823-6bb1-4e63-8729-efed71f79529" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.025882 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.072129 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ktxhf"] Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.074294 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.082973 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ktxhf"] Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.210018 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-catalog-content\") pod \"redhat-operators-ktxhf\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.210112 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-utilities\") pod \"redhat-operators-ktxhf\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.210199 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b5gf\" (UniqueName: \"kubernetes.io/projected/b684281c-a78f-4328-a682-b1571b1df08b-kube-api-access-8b5gf\") pod \"redhat-operators-ktxhf\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.316632 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-utilities\") pod \"redhat-operators-ktxhf\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.316754 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b5gf\" (UniqueName: \"kubernetes.io/projected/b684281c-a78f-4328-a682-b1571b1df08b-kube-api-access-8b5gf\") pod \"redhat-operators-ktxhf\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.316827 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-catalog-content\") pod \"redhat-operators-ktxhf\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.318064 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-catalog-content\") pod \"redhat-operators-ktxhf\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.318398 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-utilities\") pod \"redhat-operators-ktxhf\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.343845 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b5gf\" (UniqueName: \"kubernetes.io/projected/b684281c-a78f-4328-a682-b1571b1df08b-kube-api-access-8b5gf\") pod \"redhat-operators-ktxhf\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.349074 4871 generic.go:334] "Generic (PLEG): container finished" podID="34f458c3-4674-45bf-a2e6-34d35adbde8c" containerID="d830519f81b99e97eda54435bba44f2aff636773ac05de6dab11019461179590" exitCode=0 Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.349220 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"34f458c3-4674-45bf-a2e6-34d35adbde8c","Type":"ContainerDied","Data":"d830519f81b99e97eda54435bba44f2aff636773ac05de6dab11019461179590"} Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.363763 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" event={"ID":"1a916502-6bc9-48ba-906e-831e6fe85e3b","Type":"ContainerDied","Data":"d4b10e6afa8ba93e6b6f0991ec6b876659f82252aa9380d4fc4dbe33d728d029"} Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.363838 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4b10e6afa8ba93e6b6f0991ec6b876659f82252aa9380d4fc4dbe33d728d029" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.363872 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.366954 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95hj" event={"ID":"f43a1d45-76b6-4fab-aaf4-150c13787a3d","Type":"ContainerStarted","Data":"e3c670e2a236f06a736c535b12337cf7ecc4a26e99ea0f5c8324e3242861a9fe"} Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.367037 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95hj" event={"ID":"f43a1d45-76b6-4fab-aaf4-150c13787a3d","Type":"ContainerStarted","Data":"160ee2f07d3c382a75584aecef861405e727ab4e152b2baf65968629e321c020"} Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.381501 4871 generic.go:334] "Generic (PLEG): container finished" podID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerID="7f70dbc8a81ee237e3e4c88ea227ab4a60f12943ea0066a3b1a0644e941e9cc2" exitCode=0 Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.381637 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbjpj" event={"ID":"cc78d2b6-52d6-413a-9f72-5b86a26ff79a","Type":"ContainerDied","Data":"7f70dbc8a81ee237e3e4c88ea227ab4a60f12943ea0066a3b1a0644e941e9cc2"} Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.383171 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbjpj" event={"ID":"cc78d2b6-52d6-413a-9f72-5b86a26ff79a","Type":"ContainerStarted","Data":"24e785dd251e7d481f5ec326b13c3d7299477b45789f53e55646923ff1457e90"} Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.405291 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.408696 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.411493 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.424375 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.428342 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.429327 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.502485 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gjghv"] Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.518869 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.518914 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.620298 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.620345 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.620455 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.644540 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.649031 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ktxhf"] Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.737766 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.967357 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:21 crc kubenswrapper[4871]: I1007 22:11:21.976404 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-r8db4" Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.286904 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.402416 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a","Type":"ContainerStarted","Data":"634932e327df6831f3c301d9b0abecba46835592776472456ed5fce32dc91934"} Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.406817 4871 generic.go:334] "Generic (PLEG): container finished" podID="91c74861-3e58-4603-b0fd-d791b0368823" containerID="2fe8798016613c1af5ec77168244b8079b68dc7b6781bc39d087a9ec47fb19cb" exitCode=0 Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.407271 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjghv" event={"ID":"91c74861-3e58-4603-b0fd-d791b0368823","Type":"ContainerDied","Data":"2fe8798016613c1af5ec77168244b8079b68dc7b6781bc39d087a9ec47fb19cb"} Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.407342 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjghv" event={"ID":"91c74861-3e58-4603-b0fd-d791b0368823","Type":"ContainerStarted","Data":"b6c217c4bb047a1c3e9394b338af1e6b6d9366865df84bbbbee1c66d0bfacf98"} Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.411274 4871 generic.go:334] "Generic (PLEG): container finished" podID="b684281c-a78f-4328-a682-b1571b1df08b" containerID="0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b" exitCode=0 Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.411397 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktxhf" event={"ID":"b684281c-a78f-4328-a682-b1571b1df08b","Type":"ContainerDied","Data":"0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b"} Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.411468 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktxhf" event={"ID":"b684281c-a78f-4328-a682-b1571b1df08b","Type":"ContainerStarted","Data":"e9e7f12a931c688c760c43ff2a9f3172e91416ac9ca1105980fc9fb332859342"} Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.415552 4871 generic.go:334] "Generic (PLEG): container finished" podID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerID="e3c670e2a236f06a736c535b12337cf7ecc4a26e99ea0f5c8324e3242861a9fe" exitCode=0 Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.418202 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95hj" event={"ID":"f43a1d45-76b6-4fab-aaf4-150c13787a3d","Type":"ContainerDied","Data":"e3c670e2a236f06a736c535b12337cf7ecc4a26e99ea0f5c8324e3242861a9fe"} Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.747234 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.847298 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34f458c3-4674-45bf-a2e6-34d35adbde8c-kubelet-dir\") pod \"34f458c3-4674-45bf-a2e6-34d35adbde8c\" (UID: \"34f458c3-4674-45bf-a2e6-34d35adbde8c\") " Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.847367 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34f458c3-4674-45bf-a2e6-34d35adbde8c-kube-api-access\") pod \"34f458c3-4674-45bf-a2e6-34d35adbde8c\" (UID: \"34f458c3-4674-45bf-a2e6-34d35adbde8c\") " Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.847637 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34f458c3-4674-45bf-a2e6-34d35adbde8c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "34f458c3-4674-45bf-a2e6-34d35adbde8c" (UID: "34f458c3-4674-45bf-a2e6-34d35adbde8c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.847785 4871 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/34f458c3-4674-45bf-a2e6-34d35adbde8c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.863228 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f458c3-4674-45bf-a2e6-34d35adbde8c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "34f458c3-4674-45bf-a2e6-34d35adbde8c" (UID: "34f458c3-4674-45bf-a2e6-34d35adbde8c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:11:22 crc kubenswrapper[4871]: I1007 22:11:22.949753 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34f458c3-4674-45bf-a2e6-34d35adbde8c-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 22:11:23 crc kubenswrapper[4871]: I1007 22:11:23.450356 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a","Type":"ContainerStarted","Data":"a03346feb7773274e39f51c0d763e138802b7a9fb1270ca91ff864cacf595c08"} Oct 07 22:11:23 crc kubenswrapper[4871]: I1007 22:11:23.453334 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"34f458c3-4674-45bf-a2e6-34d35adbde8c","Type":"ContainerDied","Data":"de7e95af65521655c39046678555e85b06e97ac6da75fc9b5fda3a786e3df0c0"} Oct 07 22:11:23 crc kubenswrapper[4871]: I1007 22:11:23.453380 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de7e95af65521655c39046678555e85b06e97ac6da75fc9b5fda3a786e3df0c0" Oct 07 22:11:23 crc kubenswrapper[4871]: I1007 22:11:23.453439 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 22:11:23 crc kubenswrapper[4871]: I1007 22:11:23.477671 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.477642795 podStartE2EDuration="2.477642795s" podCreationTimestamp="2025-10-07 22:11:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:23.470661329 +0000 UTC m=+157.273359502" watchObservedRunningTime="2025-10-07 22:11:23.477642795 +0000 UTC m=+157.280340868" Oct 07 22:11:24 crc kubenswrapper[4871]: I1007 22:11:24.480659 4871 generic.go:334] "Generic (PLEG): container finished" podID="dee1d88c-8bd3-459c-a2e2-b87ba65fad2a" containerID="a03346feb7773274e39f51c0d763e138802b7a9fb1270ca91ff864cacf595c08" exitCode=0 Oct 07 22:11:24 crc kubenswrapper[4871]: I1007 22:11:24.480718 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a","Type":"ContainerDied","Data":"a03346feb7773274e39f51c0d763e138802b7a9fb1270ca91ff864cacf595c08"} Oct 07 22:11:25 crc kubenswrapper[4871]: I1007 22:11:25.470530 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-w49nv" Oct 07 22:11:25 crc kubenswrapper[4871]: I1007 22:11:25.824832 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:11:29 crc kubenswrapper[4871]: I1007 22:11:29.024926 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-hsm2n" Oct 07 22:11:30 crc kubenswrapper[4871]: I1007 22:11:30.255094 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:30 crc kubenswrapper[4871]: I1007 22:11:30.269999 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.035657 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.059170 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/684fd639-9051-4c1a-8bf0-9d1cb7b5cf59-metrics-certs\") pod \"network-metrics-daemon-gbxdg\" (UID: \"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59\") " pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.208733 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbxdg" Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.710073 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.768991 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kubelet-dir\") pod \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\" (UID: \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\") " Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.769141 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kube-api-access\") pod \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\" (UID: \"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a\") " Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.769145 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "dee1d88c-8bd3-459c-a2e2-b87ba65fad2a" (UID: "dee1d88c-8bd3-459c-a2e2-b87ba65fad2a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.769395 4871 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.774523 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "dee1d88c-8bd3-459c-a2e2-b87ba65fad2a" (UID: "dee1d88c-8bd3-459c-a2e2-b87ba65fad2a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:11:31 crc kubenswrapper[4871]: I1007 22:11:31.870618 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dee1d88c-8bd3-459c-a2e2-b87ba65fad2a-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 22:11:32 crc kubenswrapper[4871]: I1007 22:11:32.555312 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dee1d88c-8bd3-459c-a2e2-b87ba65fad2a","Type":"ContainerDied","Data":"634932e327df6831f3c301d9b0abecba46835592776472456ed5fce32dc91934"} Oct 07 22:11:32 crc kubenswrapper[4871]: I1007 22:11:32.555770 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="634932e327df6831f3c301d9b0abecba46835592776472456ed5fce32dc91934" Oct 07 22:11:32 crc kubenswrapper[4871]: I1007 22:11:32.555427 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 22:11:35 crc kubenswrapper[4871]: I1007 22:11:35.513021 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:11:35 crc kubenswrapper[4871]: I1007 22:11:35.513179 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:11:37 crc kubenswrapper[4871]: I1007 22:11:37.507212 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:11:50 crc kubenswrapper[4871]: I1007 22:11:50.392949 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4p54k" Oct 07 22:11:53 crc kubenswrapper[4871]: E1007 22:11:53.686090 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 07 22:11:53 crc kubenswrapper[4871]: E1007 22:11:53.686740 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8b5gf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-ktxhf_openshift-marketplace(b684281c-a78f-4328-a682-b1571b1df08b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 22:11:53 crc kubenswrapper[4871]: E1007 22:11:53.687978 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-ktxhf" podUID="b684281c-a78f-4328-a682-b1571b1df08b" Oct 07 22:11:53 crc kubenswrapper[4871]: E1007 22:11:53.722349 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 07 22:11:53 crc kubenswrapper[4871]: E1007 22:11:53.722556 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x5j22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rwkd7_openshift-marketplace(d02afe1f-19f0-448f-a9a1-0e7f19b39949): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 22:11:53 crc kubenswrapper[4871]: E1007 22:11:53.723780 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rwkd7" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" Oct 07 22:11:53 crc kubenswrapper[4871]: E1007 22:11:53.730464 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 07 22:11:53 crc kubenswrapper[4871]: E1007 22:11:53.730674 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qvqcw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-dqpjz_openshift-marketplace(d868b268-45e4-480f-b84c-b74b51127ae6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 22:11:53 crc kubenswrapper[4871]: E1007 22:11:53.731926 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-dqpjz" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.250675 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-ktxhf" podUID="b684281c-a78f-4328-a682-b1571b1df08b" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.333869 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.334899 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pdx48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-x95hj_openshift-marketplace(f43a1d45-76b6-4fab-aaf4-150c13787a3d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.336492 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-x95hj" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.346053 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.346287 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dp6kt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pbjpj_openshift-marketplace(cc78d2b6-52d6-413a-9f72-5b86a26ff79a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.347903 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pbjpj" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" Oct 07 22:11:54 crc kubenswrapper[4871]: I1007 22:11:54.494965 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gbxdg"] Oct 07 22:11:54 crc kubenswrapper[4871]: W1007 22:11:54.521476 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod684fd639_9051_4c1a_8bf0_9d1cb7b5cf59.slice/crio-1790c7d7d0782a6b601ebdd00584cf75937056f7860c51f43d8622d86eaff92a WatchSource:0}: Error finding container 1790c7d7d0782a6b601ebdd00584cf75937056f7860c51f43d8622d86eaff92a: Status 404 returned error can't find the container with id 1790c7d7d0782a6b601ebdd00584cf75937056f7860c51f43d8622d86eaff92a Oct 07 22:11:54 crc kubenswrapper[4871]: I1007 22:11:54.717473 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjghv" event={"ID":"91c74861-3e58-4603-b0fd-d791b0368823","Type":"ContainerStarted","Data":"fc429956e4c0d480fb3c46005fb0cfedaf5839d46871c246f70ee338e6baf62d"} Oct 07 22:11:54 crc kubenswrapper[4871]: I1007 22:11:54.718426 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" event={"ID":"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59","Type":"ContainerStarted","Data":"1790c7d7d0782a6b601ebdd00584cf75937056f7860c51f43d8622d86eaff92a"} Oct 07 22:11:54 crc kubenswrapper[4871]: I1007 22:11:54.721274 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dt75j" event={"ID":"813d2075-6f88-4e62-9d27-575353b29476","Type":"ContainerStarted","Data":"160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428"} Oct 07 22:11:54 crc kubenswrapper[4871]: I1007 22:11:54.722882 4871 generic.go:334] "Generic (PLEG): container finished" podID="483620a6-7917-403c-af03-8567c258e600" containerID="fb94e89d3dae8ed267d3d3ceddb9c6934fd0e118d39c716c6ec4195992f530c8" exitCode=0 Oct 07 22:11:54 crc kubenswrapper[4871]: I1007 22:11:54.723689 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hn2s5" event={"ID":"483620a6-7917-403c-af03-8567c258e600","Type":"ContainerDied","Data":"fb94e89d3dae8ed267d3d3ceddb9c6934fd0e118d39c716c6ec4195992f530c8"} Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.725133 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-pbjpj" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.725173 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-x95hj" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.726751 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rwkd7" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" Oct 07 22:11:54 crc kubenswrapper[4871]: E1007 22:11:54.738259 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-dqpjz" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" Oct 07 22:11:55 crc kubenswrapper[4871]: I1007 22:11:55.732961 4871 generic.go:334] "Generic (PLEG): container finished" podID="91c74861-3e58-4603-b0fd-d791b0368823" containerID="fc429956e4c0d480fb3c46005fb0cfedaf5839d46871c246f70ee338e6baf62d" exitCode=0 Oct 07 22:11:55 crc kubenswrapper[4871]: I1007 22:11:55.733052 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjghv" event={"ID":"91c74861-3e58-4603-b0fd-d791b0368823","Type":"ContainerDied","Data":"fc429956e4c0d480fb3c46005fb0cfedaf5839d46871c246f70ee338e6baf62d"} Oct 07 22:11:55 crc kubenswrapper[4871]: I1007 22:11:55.737964 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" event={"ID":"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59","Type":"ContainerStarted","Data":"9442fca23db70b43efbc777c7356d34737793c1f4e980751cff59cbcb8984fc3"} Oct 07 22:11:55 crc kubenswrapper[4871]: I1007 22:11:55.738076 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gbxdg" event={"ID":"684fd639-9051-4c1a-8bf0-9d1cb7b5cf59","Type":"ContainerStarted","Data":"d96dcca1f265ae932443ed196afc23b89bd12d9b052511eaed25e414b4c580b9"} Oct 07 22:11:55 crc kubenswrapper[4871]: I1007 22:11:55.740850 4871 generic.go:334] "Generic (PLEG): container finished" podID="813d2075-6f88-4e62-9d27-575353b29476" containerID="160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428" exitCode=0 Oct 07 22:11:55 crc kubenswrapper[4871]: I1007 22:11:55.740909 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dt75j" event={"ID":"813d2075-6f88-4e62-9d27-575353b29476","Type":"ContainerDied","Data":"160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428"} Oct 07 22:11:55 crc kubenswrapper[4871]: I1007 22:11:55.784545 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-gbxdg" podStartSLOduration=167.784523724 podStartE2EDuration="2m47.784523724s" podCreationTimestamp="2025-10-07 22:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:11:55.784277977 +0000 UTC m=+189.586976090" watchObservedRunningTime="2025-10-07 22:11:55.784523724 +0000 UTC m=+189.587221807" Oct 07 22:11:55 crc kubenswrapper[4871]: I1007 22:11:55.833118 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 22:11:56 crc kubenswrapper[4871]: I1007 22:11:56.749631 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dt75j" event={"ID":"813d2075-6f88-4e62-9d27-575353b29476","Type":"ContainerStarted","Data":"375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02"} Oct 07 22:11:56 crc kubenswrapper[4871]: I1007 22:11:56.753971 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hn2s5" event={"ID":"483620a6-7917-403c-af03-8567c258e600","Type":"ContainerStarted","Data":"b3facacde2046614c478416276d85ab499e2619592bac1b4f240bb12853543f1"} Oct 07 22:11:56 crc kubenswrapper[4871]: I1007 22:11:56.759236 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjghv" event={"ID":"91c74861-3e58-4603-b0fd-d791b0368823","Type":"ContainerStarted","Data":"32362cac514e2e27e1438c49e9e6a37586be3c32bb961e5261eab96d18d84a7b"} Oct 07 22:11:56 crc kubenswrapper[4871]: I1007 22:11:56.779716 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dt75j" podStartSLOduration=1.7942038139999998 podStartE2EDuration="38.779660618s" podCreationTimestamp="2025-10-07 22:11:18 +0000 UTC" firstStartedPulling="2025-10-07 22:11:19.261365753 +0000 UTC m=+153.064063826" lastFinishedPulling="2025-10-07 22:11:56.246822557 +0000 UTC m=+190.049520630" observedRunningTime="2025-10-07 22:11:56.774554692 +0000 UTC m=+190.577252775" watchObservedRunningTime="2025-10-07 22:11:56.779660618 +0000 UTC m=+190.582358691" Oct 07 22:11:56 crc kubenswrapper[4871]: I1007 22:11:56.795785 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gjghv" podStartSLOduration=2.862554621 podStartE2EDuration="36.795770208s" podCreationTimestamp="2025-10-07 22:11:20 +0000 UTC" firstStartedPulling="2025-10-07 22:11:22.426854086 +0000 UTC m=+156.229552159" lastFinishedPulling="2025-10-07 22:11:56.360069673 +0000 UTC m=+190.162767746" observedRunningTime="2025-10-07 22:11:56.795311155 +0000 UTC m=+190.598009218" watchObservedRunningTime="2025-10-07 22:11:56.795770208 +0000 UTC m=+190.598468281" Oct 07 22:11:56 crc kubenswrapper[4871]: I1007 22:11:56.816501 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hn2s5" podStartSLOduration=3.281174886 podStartE2EDuration="39.816477469s" podCreationTimestamp="2025-10-07 22:11:17 +0000 UTC" firstStartedPulling="2025-10-07 22:11:19.267537288 +0000 UTC m=+153.070235361" lastFinishedPulling="2025-10-07 22:11:55.802839831 +0000 UTC m=+189.605537944" observedRunningTime="2025-10-07 22:11:56.815332549 +0000 UTC m=+190.618030622" watchObservedRunningTime="2025-10-07 22:11:56.816477469 +0000 UTC m=+190.619175542" Oct 07 22:11:58 crc kubenswrapper[4871]: I1007 22:11:58.194271 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:58 crc kubenswrapper[4871]: I1007 22:11:58.194382 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:11:58 crc kubenswrapper[4871]: I1007 22:11:58.407764 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:58 crc kubenswrapper[4871]: I1007 22:11:58.407843 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:11:59 crc kubenswrapper[4871]: I1007 22:11:59.356099 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hn2s5" podUID="483620a6-7917-403c-af03-8567c258e600" containerName="registry-server" probeResult="failure" output=< Oct 07 22:11:59 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 22:11:59 crc kubenswrapper[4871]: > Oct 07 22:11:59 crc kubenswrapper[4871]: I1007 22:11:59.454674 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-dt75j" podUID="813d2075-6f88-4e62-9d27-575353b29476" containerName="registry-server" probeResult="failure" output=< Oct 07 22:11:59 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 22:11:59 crc kubenswrapper[4871]: > Oct 07 22:12:01 crc kubenswrapper[4871]: I1007 22:12:01.026659 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:12:01 crc kubenswrapper[4871]: I1007 22:12:01.027217 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:12:02 crc kubenswrapper[4871]: I1007 22:12:02.075967 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gjghv" podUID="91c74861-3e58-4603-b0fd-d791b0368823" containerName="registry-server" probeResult="failure" output=< Oct 07 22:12:02 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 22:12:02 crc kubenswrapper[4871]: > Oct 07 22:12:05 crc kubenswrapper[4871]: I1007 22:12:05.512225 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:12:05 crc kubenswrapper[4871]: I1007 22:12:05.512770 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:12:08 crc kubenswrapper[4871]: I1007 22:12:08.357001 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:12:08 crc kubenswrapper[4871]: I1007 22:12:08.416315 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:12:08 crc kubenswrapper[4871]: I1007 22:12:08.479108 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:12:08 crc kubenswrapper[4871]: I1007 22:12:08.537966 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:12:09 crc kubenswrapper[4871]: I1007 22:12:09.844140 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hn2s5"] Oct 07 22:12:09 crc kubenswrapper[4871]: I1007 22:12:09.853517 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hn2s5" podUID="483620a6-7917-403c-af03-8567c258e600" containerName="registry-server" containerID="cri-o://b3facacde2046614c478416276d85ab499e2619592bac1b4f240bb12853543f1" gracePeriod=2 Oct 07 22:12:10 crc kubenswrapper[4871]: I1007 22:12:10.837188 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dt75j"] Oct 07 22:12:10 crc kubenswrapper[4871]: I1007 22:12:10.838115 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dt75j" podUID="813d2075-6f88-4e62-9d27-575353b29476" containerName="registry-server" containerID="cri-o://375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02" gracePeriod=2 Oct 07 22:12:10 crc kubenswrapper[4871]: I1007 22:12:10.866036 4871 generic.go:334] "Generic (PLEG): container finished" podID="483620a6-7917-403c-af03-8567c258e600" containerID="b3facacde2046614c478416276d85ab499e2619592bac1b4f240bb12853543f1" exitCode=0 Oct 07 22:12:10 crc kubenswrapper[4871]: I1007 22:12:10.866127 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hn2s5" event={"ID":"483620a6-7917-403c-af03-8567c258e600","Type":"ContainerDied","Data":"b3facacde2046614c478416276d85ab499e2619592bac1b4f240bb12853543f1"} Oct 07 22:12:10 crc kubenswrapper[4871]: I1007 22:12:10.869322 4871 generic.go:334] "Generic (PLEG): container finished" podID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerID="68858ef6583dfb1f51ded1c27ea80c0331fdb62fc1e2ae92b0d7d3dce359f076" exitCode=0 Oct 07 22:12:10 crc kubenswrapper[4871]: I1007 22:12:10.869351 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwkd7" event={"ID":"d02afe1f-19f0-448f-a9a1-0e7f19b39949","Type":"ContainerDied","Data":"68858ef6583dfb1f51ded1c27ea80c0331fdb62fc1e2ae92b0d7d3dce359f076"} Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.092279 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.147223 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.294429 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.367338 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.438021 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-utilities\") pod \"483620a6-7917-403c-af03-8567c258e600\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.438423 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzqqc\" (UniqueName: \"kubernetes.io/projected/813d2075-6f88-4e62-9d27-575353b29476-kube-api-access-gzqqc\") pod \"813d2075-6f88-4e62-9d27-575353b29476\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.438563 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-catalog-content\") pod \"483620a6-7917-403c-af03-8567c258e600\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.438722 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-utilities\") pod \"813d2075-6f88-4e62-9d27-575353b29476\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.438856 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p94kn\" (UniqueName: \"kubernetes.io/projected/483620a6-7917-403c-af03-8567c258e600-kube-api-access-p94kn\") pod \"483620a6-7917-403c-af03-8567c258e600\" (UID: \"483620a6-7917-403c-af03-8567c258e600\") " Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.438980 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-catalog-content\") pod \"813d2075-6f88-4e62-9d27-575353b29476\" (UID: \"813d2075-6f88-4e62-9d27-575353b29476\") " Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.447274 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/483620a6-7917-403c-af03-8567c258e600-kube-api-access-p94kn" (OuterVolumeSpecName: "kube-api-access-p94kn") pod "483620a6-7917-403c-af03-8567c258e600" (UID: "483620a6-7917-403c-af03-8567c258e600"). InnerVolumeSpecName "kube-api-access-p94kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.456372 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-utilities" (OuterVolumeSpecName: "utilities") pod "483620a6-7917-403c-af03-8567c258e600" (UID: "483620a6-7917-403c-af03-8567c258e600"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.457094 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/813d2075-6f88-4e62-9d27-575353b29476-kube-api-access-gzqqc" (OuterVolumeSpecName: "kube-api-access-gzqqc") pod "813d2075-6f88-4e62-9d27-575353b29476" (UID: "813d2075-6f88-4e62-9d27-575353b29476"). InnerVolumeSpecName "kube-api-access-gzqqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.457561 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-utilities" (OuterVolumeSpecName: "utilities") pod "813d2075-6f88-4e62-9d27-575353b29476" (UID: "813d2075-6f88-4e62-9d27-575353b29476"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.506773 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "483620a6-7917-403c-af03-8567c258e600" (UID: "483620a6-7917-403c-af03-8567c258e600"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.512976 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "813d2075-6f88-4e62-9d27-575353b29476" (UID: "813d2075-6f88-4e62-9d27-575353b29476"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.540778 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.540823 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.540835 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p94kn\" (UniqueName: \"kubernetes.io/projected/483620a6-7917-403c-af03-8567c258e600-kube-api-access-p94kn\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.540847 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/813d2075-6f88-4e62-9d27-575353b29476-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.540856 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483620a6-7917-403c-af03-8567c258e600-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.540866 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzqqc\" (UniqueName: \"kubernetes.io/projected/813d2075-6f88-4e62-9d27-575353b29476-kube-api-access-gzqqc\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.878075 4871 generic.go:334] "Generic (PLEG): container finished" podID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerID="43867ff257f30b4e76103c0af651971edc0c66a7290985354bddf8252e2dc22c" exitCode=0 Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.878186 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbjpj" event={"ID":"cc78d2b6-52d6-413a-9f72-5b86a26ff79a","Type":"ContainerDied","Data":"43867ff257f30b4e76103c0af651971edc0c66a7290985354bddf8252e2dc22c"} Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.883522 4871 generic.go:334] "Generic (PLEG): container finished" podID="813d2075-6f88-4e62-9d27-575353b29476" containerID="375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02" exitCode=0 Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.883650 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dt75j" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.883682 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dt75j" event={"ID":"813d2075-6f88-4e62-9d27-575353b29476","Type":"ContainerDied","Data":"375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02"} Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.883738 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dt75j" event={"ID":"813d2075-6f88-4e62-9d27-575353b29476","Type":"ContainerDied","Data":"013e60ae5f426de2054ba1be4cb236498ad533aaa7607b24243bc7409142ed5c"} Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.883768 4871 scope.go:117] "RemoveContainer" containerID="375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.886676 4871 generic.go:334] "Generic (PLEG): container finished" podID="d868b268-45e4-480f-b84c-b74b51127ae6" containerID="1d1fb46f483f549bdaec907aa5f675dfe36ec10c80a9c2041f6f83a9e958fad8" exitCode=0 Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.886746 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqpjz" event={"ID":"d868b268-45e4-480f-b84c-b74b51127ae6","Type":"ContainerDied","Data":"1d1fb46f483f549bdaec907aa5f675dfe36ec10c80a9c2041f6f83a9e958fad8"} Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.892912 4871 generic.go:334] "Generic (PLEG): container finished" podID="b684281c-a78f-4328-a682-b1571b1df08b" containerID="4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de" exitCode=0 Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.892990 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktxhf" event={"ID":"b684281c-a78f-4328-a682-b1571b1df08b","Type":"ContainerDied","Data":"4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de"} Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.904708 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hn2s5" event={"ID":"483620a6-7917-403c-af03-8567c258e600","Type":"ContainerDied","Data":"d9a1f827a3f12a1bc704e5b9853d2f445a3e237453c83a724b34dd8bf6850791"} Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.904912 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hn2s5" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.905705 4871 scope.go:117] "RemoveContainer" containerID="160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.908920 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwkd7" event={"ID":"d02afe1f-19f0-448f-a9a1-0e7f19b39949","Type":"ContainerStarted","Data":"339d971690348c1a92f2ce56fcd9a693d20a3877162dd0664f0b9529b2a560fd"} Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.914623 4871 generic.go:334] "Generic (PLEG): container finished" podID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerID="8db0330521660b534065c10d7093410f1a9bd5178e5b6c7f00db3b09783698bf" exitCode=0 Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.914735 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95hj" event={"ID":"f43a1d45-76b6-4fab-aaf4-150c13787a3d","Type":"ContainerDied","Data":"8db0330521660b534065c10d7093410f1a9bd5178e5b6c7f00db3b09783698bf"} Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.945303 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rwkd7" podStartSLOduration=2.684234824 podStartE2EDuration="54.945275745s" podCreationTimestamp="2025-10-07 22:11:17 +0000 UTC" firstStartedPulling="2025-10-07 22:11:19.275377176 +0000 UTC m=+153.078075249" lastFinishedPulling="2025-10-07 22:12:11.536418087 +0000 UTC m=+205.339116170" observedRunningTime="2025-10-07 22:12:11.93547098 +0000 UTC m=+205.738169043" watchObservedRunningTime="2025-10-07 22:12:11.945275745 +0000 UTC m=+205.747973828" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.950219 4871 scope.go:117] "RemoveContainer" containerID="f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.973712 4871 scope.go:117] "RemoveContainer" containerID="375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02" Oct 07 22:12:11 crc kubenswrapper[4871]: E1007 22:12:11.974493 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02\": container with ID starting with 375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02 not found: ID does not exist" containerID="375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.974539 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02"} err="failed to get container status \"375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02\": rpc error: code = NotFound desc = could not find container \"375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02\": container with ID starting with 375e4326922b4a150b6fd2aa5cceb541e77de94d70fb1c6f87649db2a4f2ef02 not found: ID does not exist" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.974602 4871 scope.go:117] "RemoveContainer" containerID="160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428" Oct 07 22:12:11 crc kubenswrapper[4871]: E1007 22:12:11.974943 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428\": container with ID starting with 160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428 not found: ID does not exist" containerID="160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.974969 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428"} err="failed to get container status \"160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428\": rpc error: code = NotFound desc = could not find container \"160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428\": container with ID starting with 160da2a2401204b385751a198db53ad9e9b40af06eea0b85e1733e667d957428 not found: ID does not exist" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.974986 4871 scope.go:117] "RemoveContainer" containerID="f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa" Oct 07 22:12:11 crc kubenswrapper[4871]: E1007 22:12:11.975481 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa\": container with ID starting with f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa not found: ID does not exist" containerID="f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.975505 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa"} err="failed to get container status \"f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa\": rpc error: code = NotFound desc = could not find container \"f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa\": container with ID starting with f437d75532904020bf3e97793ff3eeab4c983f5745f47e7db22a923755c07afa not found: ID does not exist" Oct 07 22:12:11 crc kubenswrapper[4871]: I1007 22:12:11.975520 4871 scope.go:117] "RemoveContainer" containerID="b3facacde2046614c478416276d85ab499e2619592bac1b4f240bb12853543f1" Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.007681 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dt75j"] Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.015847 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dt75j"] Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.019607 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hn2s5"] Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.025107 4871 scope.go:117] "RemoveContainer" containerID="fb94e89d3dae8ed267d3d3ceddb9c6934fd0e118d39c716c6ec4195992f530c8" Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.032393 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hn2s5"] Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.053217 4871 scope.go:117] "RemoveContainer" containerID="c8d18ef5598fc8b50dda841bef89d5bfaeab9102bc3274346f6f6a103efdc365" Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.924437 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqpjz" event={"ID":"d868b268-45e4-480f-b84c-b74b51127ae6","Type":"ContainerStarted","Data":"d7cca54b452a1f0159c8409f59e9da4cf5149f67361e1d0c8ccffad6de9adf27"} Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.928617 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95hj" event={"ID":"f43a1d45-76b6-4fab-aaf4-150c13787a3d","Type":"ContainerStarted","Data":"39ee471cb6b6a991211139adcf08b8ee2d50beef67e438b97a71f32d92478d0f"} Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.934449 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbjpj" event={"ID":"cc78d2b6-52d6-413a-9f72-5b86a26ff79a","Type":"ContainerStarted","Data":"4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5"} Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.947736 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dqpjz" podStartSLOduration=2.552178864 podStartE2EDuration="55.947713528s" podCreationTimestamp="2025-10-07 22:11:17 +0000 UTC" firstStartedPulling="2025-10-07 22:11:19.264899047 +0000 UTC m=+153.067597120" lastFinishedPulling="2025-10-07 22:12:12.660433721 +0000 UTC m=+206.463131784" observedRunningTime="2025-10-07 22:12:12.947507092 +0000 UTC m=+206.750205165" watchObservedRunningTime="2025-10-07 22:12:12.947713528 +0000 UTC m=+206.750411601" Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.987985 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x95hj" podStartSLOduration=2.87915077 podStartE2EDuration="52.987964794s" podCreationTimestamp="2025-10-07 22:11:20 +0000 UTC" firstStartedPulling="2025-10-07 22:11:22.426834786 +0000 UTC m=+156.229532859" lastFinishedPulling="2025-10-07 22:12:12.53564881 +0000 UTC m=+206.338346883" observedRunningTime="2025-10-07 22:12:12.978020665 +0000 UTC m=+206.780718728" watchObservedRunningTime="2025-10-07 22:12:12.987964794 +0000 UTC m=+206.790662867" Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.995092 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="483620a6-7917-403c-af03-8567c258e600" path="/var/lib/kubelet/pods/483620a6-7917-403c-af03-8567c258e600/volumes" Oct 07 22:12:12 crc kubenswrapper[4871]: I1007 22:12:12.995752 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="813d2075-6f88-4e62-9d27-575353b29476" path="/var/lib/kubelet/pods/813d2075-6f88-4e62-9d27-575353b29476/volumes" Oct 07 22:12:13 crc kubenswrapper[4871]: I1007 22:12:13.004600 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pbjpj" podStartSLOduration=2.919927895 podStartE2EDuration="54.004572058s" podCreationTimestamp="2025-10-07 22:11:19 +0000 UTC" firstStartedPulling="2025-10-07 22:11:21.389029613 +0000 UTC m=+155.191727686" lastFinishedPulling="2025-10-07 22:12:12.473673756 +0000 UTC m=+206.276371849" observedRunningTime="2025-10-07 22:12:13.003798167 +0000 UTC m=+206.806496240" watchObservedRunningTime="2025-10-07 22:12:13.004572058 +0000 UTC m=+206.807270121" Oct 07 22:12:13 crc kubenswrapper[4871]: I1007 22:12:13.950460 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktxhf" event={"ID":"b684281c-a78f-4328-a682-b1571b1df08b","Type":"ContainerStarted","Data":"bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac"} Oct 07 22:12:13 crc kubenswrapper[4871]: I1007 22:12:13.988073 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ktxhf" podStartSLOduration=2.580747606 podStartE2EDuration="52.988050441s" podCreationTimestamp="2025-10-07 22:11:21 +0000 UTC" firstStartedPulling="2025-10-07 22:11:22.427222506 +0000 UTC m=+156.229920579" lastFinishedPulling="2025-10-07 22:12:12.834525341 +0000 UTC m=+206.637223414" observedRunningTime="2025-10-07 22:12:13.982865276 +0000 UTC m=+207.785563369" watchObservedRunningTime="2025-10-07 22:12:13.988050441 +0000 UTC m=+207.790748524" Oct 07 22:12:17 crc kubenswrapper[4871]: I1007 22:12:17.785143 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:12:17 crc kubenswrapper[4871]: I1007 22:12:17.786124 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:12:17 crc kubenswrapper[4871]: I1007 22:12:17.844152 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:12:18 crc kubenswrapper[4871]: I1007 22:12:18.005493 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:12:18 crc kubenswrapper[4871]: I1007 22:12:18.006214 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:12:18 crc kubenswrapper[4871]: I1007 22:12:18.024556 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:12:18 crc kubenswrapper[4871]: I1007 22:12:18.051614 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:12:19 crc kubenswrapper[4871]: I1007 22:12:19.038538 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:12:19 crc kubenswrapper[4871]: I1007 22:12:19.983962 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:12:19 crc kubenswrapper[4871]: I1007 22:12:19.984012 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:12:20 crc kubenswrapper[4871]: I1007 22:12:20.028154 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:12:20 crc kubenswrapper[4871]: I1007 22:12:20.073265 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:12:20 crc kubenswrapper[4871]: I1007 22:12:20.416856 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:12:20 crc kubenswrapper[4871]: I1007 22:12:20.416925 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:12:20 crc kubenswrapper[4871]: I1007 22:12:20.462935 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:12:21 crc kubenswrapper[4871]: I1007 22:12:21.037294 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:12:21 crc kubenswrapper[4871]: I1007 22:12:21.405617 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:12:21 crc kubenswrapper[4871]: I1007 22:12:21.405710 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:12:21 crc kubenswrapper[4871]: I1007 22:12:21.457052 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:12:22 crc kubenswrapper[4871]: I1007 22:12:22.052174 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:12:23 crc kubenswrapper[4871]: I1007 22:12:23.235038 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95hj"] Oct 07 22:12:23 crc kubenswrapper[4871]: I1007 22:12:23.235357 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x95hj" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerName="registry-server" containerID="cri-o://39ee471cb6b6a991211139adcf08b8ee2d50beef67e438b97a71f32d92478d0f" gracePeriod=2 Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.018344 4871 generic.go:334] "Generic (PLEG): container finished" podID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerID="39ee471cb6b6a991211139adcf08b8ee2d50beef67e438b97a71f32d92478d0f" exitCode=0 Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.018430 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95hj" event={"ID":"f43a1d45-76b6-4fab-aaf4-150c13787a3d","Type":"ContainerDied","Data":"39ee471cb6b6a991211139adcf08b8ee2d50beef67e438b97a71f32d92478d0f"} Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.182383 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.344542 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-catalog-content\") pod \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.344707 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdx48\" (UniqueName: \"kubernetes.io/projected/f43a1d45-76b6-4fab-aaf4-150c13787a3d-kube-api-access-pdx48\") pod \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.344911 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-utilities\") pod \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\" (UID: \"f43a1d45-76b6-4fab-aaf4-150c13787a3d\") " Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.346188 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-utilities" (OuterVolumeSpecName: "utilities") pod "f43a1d45-76b6-4fab-aaf4-150c13787a3d" (UID: "f43a1d45-76b6-4fab-aaf4-150c13787a3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.353144 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f43a1d45-76b6-4fab-aaf4-150c13787a3d-kube-api-access-pdx48" (OuterVolumeSpecName: "kube-api-access-pdx48") pod "f43a1d45-76b6-4fab-aaf4-150c13787a3d" (UID: "f43a1d45-76b6-4fab-aaf4-150c13787a3d"). InnerVolumeSpecName "kube-api-access-pdx48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.362959 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f43a1d45-76b6-4fab-aaf4-150c13787a3d" (UID: "f43a1d45-76b6-4fab-aaf4-150c13787a3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.446843 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.447190 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43a1d45-76b6-4fab-aaf4-150c13787a3d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:24 crc kubenswrapper[4871]: I1007 22:12:24.447206 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdx48\" (UniqueName: \"kubernetes.io/projected/f43a1d45-76b6-4fab-aaf4-150c13787a3d-kube-api-access-pdx48\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:25 crc kubenswrapper[4871]: I1007 22:12:25.039740 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95hj" event={"ID":"f43a1d45-76b6-4fab-aaf4-150c13787a3d","Type":"ContainerDied","Data":"160ee2f07d3c382a75584aecef861405e727ab4e152b2baf65968629e321c020"} Oct 07 22:12:25 crc kubenswrapper[4871]: I1007 22:12:25.039850 4871 scope.go:117] "RemoveContainer" containerID="39ee471cb6b6a991211139adcf08b8ee2d50beef67e438b97a71f32d92478d0f" Oct 07 22:12:25 crc kubenswrapper[4871]: I1007 22:12:25.040109 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x95hj" Oct 07 22:12:25 crc kubenswrapper[4871]: I1007 22:12:25.062177 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95hj"] Oct 07 22:12:25 crc kubenswrapper[4871]: I1007 22:12:25.066174 4871 scope.go:117] "RemoveContainer" containerID="8db0330521660b534065c10d7093410f1a9bd5178e5b6c7f00db3b09783698bf" Oct 07 22:12:25 crc kubenswrapper[4871]: I1007 22:12:25.069075 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95hj"] Oct 07 22:12:25 crc kubenswrapper[4871]: I1007 22:12:25.083390 4871 scope.go:117] "RemoveContainer" containerID="e3c670e2a236f06a736c535b12337cf7ecc4a26e99ea0f5c8324e3242861a9fe" Oct 07 22:12:25 crc kubenswrapper[4871]: I1007 22:12:25.631872 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ktxhf"] Oct 07 22:12:25 crc kubenswrapper[4871]: I1007 22:12:25.632144 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ktxhf" podUID="b684281c-a78f-4328-a682-b1571b1df08b" containerName="registry-server" containerID="cri-o://bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac" gracePeriod=2 Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.022271 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.048060 4871 generic.go:334] "Generic (PLEG): container finished" podID="b684281c-a78f-4328-a682-b1571b1df08b" containerID="bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac" exitCode=0 Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.048129 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktxhf" event={"ID":"b684281c-a78f-4328-a682-b1571b1df08b","Type":"ContainerDied","Data":"bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac"} Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.048185 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ktxhf" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.048207 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ktxhf" event={"ID":"b684281c-a78f-4328-a682-b1571b1df08b","Type":"ContainerDied","Data":"e9e7f12a931c688c760c43ff2a9f3172e91416ac9ca1105980fc9fb332859342"} Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.048234 4871 scope.go:117] "RemoveContainer" containerID="bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.075681 4871 scope.go:117] "RemoveContainer" containerID="4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.108996 4871 scope.go:117] "RemoveContainer" containerID="0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.123688 4871 scope.go:117] "RemoveContainer" containerID="bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac" Oct 07 22:12:26 crc kubenswrapper[4871]: E1007 22:12:26.128922 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac\": container with ID starting with bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac not found: ID does not exist" containerID="bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.128967 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac"} err="failed to get container status \"bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac\": rpc error: code = NotFound desc = could not find container \"bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac\": container with ID starting with bdd20a057d29c4c425038e39371f40d720ff5122ac42e562d54a6d74ab353cac not found: ID does not exist" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.128995 4871 scope.go:117] "RemoveContainer" containerID="4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de" Oct 07 22:12:26 crc kubenswrapper[4871]: E1007 22:12:26.129331 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de\": container with ID starting with 4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de not found: ID does not exist" containerID="4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.129363 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de"} err="failed to get container status \"4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de\": rpc error: code = NotFound desc = could not find container \"4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de\": container with ID starting with 4de0c1884878c44db1efe2f813f421b5d41998aea0f19771e84b822f4189a2de not found: ID does not exist" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.129379 4871 scope.go:117] "RemoveContainer" containerID="0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b" Oct 07 22:12:26 crc kubenswrapper[4871]: E1007 22:12:26.129768 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b\": container with ID starting with 0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b not found: ID does not exist" containerID="0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.129848 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b"} err="failed to get container status \"0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b\": rpc error: code = NotFound desc = could not find container \"0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b\": container with ID starting with 0f369d1cfaf48cca8bdf68fbf087494ef8eb2a21d1e424882d2e46c2da557b8b not found: ID does not exist" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.175064 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-utilities\") pod \"b684281c-a78f-4328-a682-b1571b1df08b\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.175153 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b5gf\" (UniqueName: \"kubernetes.io/projected/b684281c-a78f-4328-a682-b1571b1df08b-kube-api-access-8b5gf\") pod \"b684281c-a78f-4328-a682-b1571b1df08b\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.175243 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-catalog-content\") pod \"b684281c-a78f-4328-a682-b1571b1df08b\" (UID: \"b684281c-a78f-4328-a682-b1571b1df08b\") " Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.176325 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-utilities" (OuterVolumeSpecName: "utilities") pod "b684281c-a78f-4328-a682-b1571b1df08b" (UID: "b684281c-a78f-4328-a682-b1571b1df08b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.183635 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b684281c-a78f-4328-a682-b1571b1df08b-kube-api-access-8b5gf" (OuterVolumeSpecName: "kube-api-access-8b5gf") pod "b684281c-a78f-4328-a682-b1571b1df08b" (UID: "b684281c-a78f-4328-a682-b1571b1df08b"). InnerVolumeSpecName "kube-api-access-8b5gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.272892 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b684281c-a78f-4328-a682-b1571b1df08b" (UID: "b684281c-a78f-4328-a682-b1571b1df08b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.277182 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b5gf\" (UniqueName: \"kubernetes.io/projected/b684281c-a78f-4328-a682-b1571b1df08b-kube-api-access-8b5gf\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.277252 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.277268 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684281c-a78f-4328-a682-b1571b1df08b-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.377732 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ktxhf"] Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.380138 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ktxhf"] Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.988849 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b684281c-a78f-4328-a682-b1571b1df08b" path="/var/lib/kubelet/pods/b684281c-a78f-4328-a682-b1571b1df08b/volumes" Oct 07 22:12:26 crc kubenswrapper[4871]: I1007 22:12:26.989642 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" path="/var/lib/kubelet/pods/f43a1d45-76b6-4fab-aaf4-150c13787a3d/volumes" Oct 07 22:12:28 crc kubenswrapper[4871]: I1007 22:12:28.864049 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-clflh"] Oct 07 22:12:35 crc kubenswrapper[4871]: I1007 22:12:35.512348 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:12:35 crc kubenswrapper[4871]: I1007 22:12:35.513109 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:12:35 crc kubenswrapper[4871]: I1007 22:12:35.513184 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:12:35 crc kubenswrapper[4871]: I1007 22:12:35.514017 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:12:35 crc kubenswrapper[4871]: I1007 22:12:35.514097 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002" gracePeriod=600 Oct 07 22:12:36 crc kubenswrapper[4871]: I1007 22:12:36.112876 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002" exitCode=0 Oct 07 22:12:36 crc kubenswrapper[4871]: I1007 22:12:36.112992 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002"} Oct 07 22:12:36 crc kubenswrapper[4871]: I1007 22:12:36.113926 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"9e9168b99e93de9001a9d600cd6b89e0f408725c87f4710f2d8b93ae3777f08a"} Oct 07 22:12:53 crc kubenswrapper[4871]: I1007 22:12:53.900338 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" podUID="f031a182-f40f-4369-83fa-51ba2f2ebd83" containerName="oauth-openshift" containerID="cri-o://d7d9d409e1bd4247dab57ab86ef337d65c1f6e7250d82dfbbf36d05a384f2500" gracePeriod=15 Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.248972 4871 generic.go:334] "Generic (PLEG): container finished" podID="f031a182-f40f-4369-83fa-51ba2f2ebd83" containerID="d7d9d409e1bd4247dab57ab86ef337d65c1f6e7250d82dfbbf36d05a384f2500" exitCode=0 Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.249093 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" event={"ID":"f031a182-f40f-4369-83fa-51ba2f2ebd83","Type":"ContainerDied","Data":"d7d9d409e1bd4247dab57ab86ef337d65c1f6e7250d82dfbbf36d05a384f2500"} Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.328048 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.386708 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr"] Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387109 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483620a6-7917-403c-af03-8567c258e600" containerName="extract-content" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387145 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="483620a6-7917-403c-af03-8567c258e600" containerName="extract-content" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387168 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b684281c-a78f-4328-a682-b1571b1df08b" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387181 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b684281c-a78f-4328-a682-b1571b1df08b" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387195 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b684281c-a78f-4328-a682-b1571b1df08b" containerName="extract-content" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387208 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b684281c-a78f-4328-a682-b1571b1df08b" containerName="extract-content" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387224 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerName="extract-utilities" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387237 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerName="extract-utilities" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387255 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483620a6-7917-403c-af03-8567c258e600" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387267 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="483620a6-7917-403c-af03-8567c258e600" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387285 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f458c3-4674-45bf-a2e6-34d35adbde8c" containerName="pruner" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387297 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f458c3-4674-45bf-a2e6-34d35adbde8c" containerName="pruner" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387312 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dee1d88c-8bd3-459c-a2e2-b87ba65fad2a" containerName="pruner" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387323 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dee1d88c-8bd3-459c-a2e2-b87ba65fad2a" containerName="pruner" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387344 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f031a182-f40f-4369-83fa-51ba2f2ebd83" containerName="oauth-openshift" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387355 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f031a182-f40f-4369-83fa-51ba2f2ebd83" containerName="oauth-openshift" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387374 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387386 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387400 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="813d2075-6f88-4e62-9d27-575353b29476" containerName="extract-utilities" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387413 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="813d2075-6f88-4e62-9d27-575353b29476" containerName="extract-utilities" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387433 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerName="extract-content" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387447 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerName="extract-content" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387467 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483620a6-7917-403c-af03-8567c258e600" containerName="extract-utilities" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387479 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="483620a6-7917-403c-af03-8567c258e600" containerName="extract-utilities" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387499 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="813d2075-6f88-4e62-9d27-575353b29476" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387511 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="813d2075-6f88-4e62-9d27-575353b29476" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387530 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b684281c-a78f-4328-a682-b1571b1df08b" containerName="extract-utilities" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387543 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b684281c-a78f-4328-a682-b1571b1df08b" containerName="extract-utilities" Oct 07 22:12:54 crc kubenswrapper[4871]: E1007 22:12:54.387555 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="813d2075-6f88-4e62-9d27-575353b29476" containerName="extract-content" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387567 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="813d2075-6f88-4e62-9d27-575353b29476" containerName="extract-content" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387737 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f031a182-f40f-4369-83fa-51ba2f2ebd83" containerName="oauth-openshift" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387757 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f458c3-4674-45bf-a2e6-34d35adbde8c" containerName="pruner" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387775 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="dee1d88c-8bd3-459c-a2e2-b87ba65fad2a" containerName="pruner" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387820 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43a1d45-76b6-4fab-aaf4-150c13787a3d" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387845 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="483620a6-7917-403c-af03-8567c258e600" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387861 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="813d2075-6f88-4e62-9d27-575353b29476" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.387881 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b684281c-a78f-4328-a682-b1571b1df08b" containerName="registry-server" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.388537 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.403700 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr"] Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428427 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-idp-0-file-data\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428478 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-dir\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428544 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-login\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428582 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-router-certs\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428630 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-service-ca\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428684 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-error\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428710 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-serving-cert\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428769 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-provider-selection\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428856 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-ocp-branding-template\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428920 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-cliconfig\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.428965 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-policies\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429023 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-trusted-ca-bundle\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429074 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-session\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429107 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6558\" (UniqueName: \"kubernetes.io/projected/f031a182-f40f-4369-83fa-51ba2f2ebd83-kube-api-access-f6558\") pod \"f031a182-f40f-4369-83fa-51ba2f2ebd83\" (UID: \"f031a182-f40f-4369-83fa-51ba2f2ebd83\") " Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429363 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429404 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-template-login\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429433 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429457 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429495 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-service-ca\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429530 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429559 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-session\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429588 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429617 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-template-error\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429659 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-router-certs\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429686 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq9c8\" (UniqueName: \"kubernetes.io/projected/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-kube-api-access-lq9c8\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429716 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-audit-policies\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429742 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-audit-dir\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.429766 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.430166 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.431572 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.432265 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.432578 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.438783 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.440550 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.442185 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.443181 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.448410 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f031a182-f40f-4369-83fa-51ba2f2ebd83-kube-api-access-f6558" (OuterVolumeSpecName: "kube-api-access-f6558") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "kube-api-access-f6558". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.448646 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.450452 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.453881 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.454935 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.464192 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f031a182-f40f-4369-83fa-51ba2f2ebd83" (UID: "f031a182-f40f-4369-83fa-51ba2f2ebd83"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.531068 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-service-ca\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.531235 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.531292 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-session\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.531331 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532216 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-service-ca\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532284 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-template-error\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532354 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-router-certs\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532510 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq9c8\" (UniqueName: \"kubernetes.io/projected/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-kube-api-access-lq9c8\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532561 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-audit-policies\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532613 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-audit-dir\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532645 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532755 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532839 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-template-login\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532878 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.532907 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533025 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533041 4871 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533058 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533080 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533096 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6558\" (UniqueName: \"kubernetes.io/projected/f031a182-f40f-4369-83fa-51ba2f2ebd83-kube-api-access-f6558\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533114 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533128 4871 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f031a182-f40f-4369-83fa-51ba2f2ebd83-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533144 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533158 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533173 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533187 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533202 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533218 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533235 4871 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f031a182-f40f-4369-83fa-51ba2f2ebd83-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533736 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-audit-dir\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.533893 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.534212 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.534327 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-audit-policies\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.538512 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-template-login\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.538522 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.539242 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-router-certs\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.539511 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.540049 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-session\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.541325 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.542709 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.544275 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-v4-0-config-user-template-error\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.564582 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq9c8\" (UniqueName: \"kubernetes.io/projected/9b663e67-560d-4d1e-9f4d-4cd71ab51cb4-kube-api-access-lq9c8\") pod \"oauth-openshift-68b6dd9b65-f6dfr\" (UID: \"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4\") " pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.711269 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:54 crc kubenswrapper[4871]: I1007 22:12:54.980652 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr"] Oct 07 22:12:55 crc kubenswrapper[4871]: I1007 22:12:55.259625 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" event={"ID":"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4","Type":"ContainerStarted","Data":"94d932a8aac75ee543c6eb52183ec36c6413157aab9bc7fe1669d8aa605fd91d"} Oct 07 22:12:55 crc kubenswrapper[4871]: I1007 22:12:55.263171 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" event={"ID":"f031a182-f40f-4369-83fa-51ba2f2ebd83","Type":"ContainerDied","Data":"5a090c0ebb6c4cbaee00164f58f606ede634f61a25768c1569345fd534eae44a"} Oct 07 22:12:55 crc kubenswrapper[4871]: I1007 22:12:55.263217 4871 scope.go:117] "RemoveContainer" containerID="d7d9d409e1bd4247dab57ab86ef337d65c1f6e7250d82dfbbf36d05a384f2500" Oct 07 22:12:55 crc kubenswrapper[4871]: I1007 22:12:55.263383 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-clflh" Oct 07 22:12:55 crc kubenswrapper[4871]: I1007 22:12:55.300578 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-clflh"] Oct 07 22:12:55 crc kubenswrapper[4871]: I1007 22:12:55.308757 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-clflh"] Oct 07 22:12:56 crc kubenswrapper[4871]: I1007 22:12:56.278830 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" event={"ID":"9b663e67-560d-4d1e-9f4d-4cd71ab51cb4","Type":"ContainerStarted","Data":"8cd92a02b7c35689f301aec48996194a4febca85b4d1149fd81d83ff95e92c93"} Oct 07 22:12:56 crc kubenswrapper[4871]: I1007 22:12:56.279307 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:56 crc kubenswrapper[4871]: I1007 22:12:56.288353 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" Oct 07 22:12:56 crc kubenswrapper[4871]: I1007 22:12:56.330515 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-68b6dd9b65-f6dfr" podStartSLOduration=28.330480176000002 podStartE2EDuration="28.330480176s" podCreationTimestamp="2025-10-07 22:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:12:56.313730188 +0000 UTC m=+250.116428311" watchObservedRunningTime="2025-10-07 22:12:56.330480176 +0000 UTC m=+250.133178289" Oct 07 22:12:56 crc kubenswrapper[4871]: I1007 22:12:56.994550 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f031a182-f40f-4369-83fa-51ba2f2ebd83" path="/var/lib/kubelet/pods/f031a182-f40f-4369-83fa-51ba2f2ebd83/volumes" Oct 07 22:14:35 crc kubenswrapper[4871]: I1007 22:14:35.512334 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:14:35 crc kubenswrapper[4871]: I1007 22:14:35.513226 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.879171 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rwkd7"] Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.880343 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rwkd7" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerName="registry-server" containerID="cri-o://339d971690348c1a92f2ce56fcd9a693d20a3877162dd0664f0b9529b2a560fd" gracePeriod=30 Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.883206 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dqpjz"] Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.883502 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dqpjz" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" containerName="registry-server" containerID="cri-o://d7cca54b452a1f0159c8409f59e9da4cf5149f67361e1d0c8ccffad6de9adf27" gracePeriod=30 Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.892308 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kvwx2"] Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.892613 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" podUID="9d97e32b-02b4-4435-bd25-81174b3e0da9" containerName="marketplace-operator" containerID="cri-o://69605fa35c1363bbb7197afd08ab593a01b606b05174b53b3f359e14c60b2cd3" gracePeriod=30 Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.902522 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbjpj"] Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.902902 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pbjpj" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerName="registry-server" containerID="cri-o://4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5" gracePeriod=30 Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.922435 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gjghv"] Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.922836 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gjghv" podUID="91c74861-3e58-4603-b0fd-d791b0368823" containerName="registry-server" containerID="cri-o://32362cac514e2e27e1438c49e9e6a37586be3c32bb961e5261eab96d18d84a7b" gracePeriod=30 Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.927975 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kgzf8"] Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.929196 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:39 crc kubenswrapper[4871]: I1007 22:14:39.937036 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kgzf8"] Oct 07 22:14:39 crc kubenswrapper[4871]: E1007 22:14:39.984688 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5 is running failed: container process not found" containerID="4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5" cmd=["grpc_health_probe","-addr=:50051"] Oct 07 22:14:39 crc kubenswrapper[4871]: E1007 22:14:39.986014 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5 is running failed: container process not found" containerID="4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5" cmd=["grpc_health_probe","-addr=:50051"] Oct 07 22:14:39 crc kubenswrapper[4871]: E1007 22:14:39.988177 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5 is running failed: container process not found" containerID="4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5" cmd=["grpc_health_probe","-addr=:50051"] Oct 07 22:14:39 crc kubenswrapper[4871]: E1007 22:14:39.988242 4871 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-pbjpj" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerName="registry-server" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.103265 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmnvw\" (UniqueName: \"kubernetes.io/projected/464efd4a-2401-457e-bf72-41ce7b9f2cac-kube-api-access-lmnvw\") pod \"marketplace-operator-79b997595-kgzf8\" (UID: \"464efd4a-2401-457e-bf72-41ce7b9f2cac\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.104960 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/464efd4a-2401-457e-bf72-41ce7b9f2cac-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kgzf8\" (UID: \"464efd4a-2401-457e-bf72-41ce7b9f2cac\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.105599 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/464efd4a-2401-457e-bf72-41ce7b9f2cac-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kgzf8\" (UID: \"464efd4a-2401-457e-bf72-41ce7b9f2cac\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.126238 4871 generic.go:334] "Generic (PLEG): container finished" podID="d868b268-45e4-480f-b84c-b74b51127ae6" containerID="d7cca54b452a1f0159c8409f59e9da4cf5149f67361e1d0c8ccffad6de9adf27" exitCode=0 Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.126351 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqpjz" event={"ID":"d868b268-45e4-480f-b84c-b74b51127ae6","Type":"ContainerDied","Data":"d7cca54b452a1f0159c8409f59e9da4cf5149f67361e1d0c8ccffad6de9adf27"} Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.133131 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwkd7" event={"ID":"d02afe1f-19f0-448f-a9a1-0e7f19b39949","Type":"ContainerDied","Data":"339d971690348c1a92f2ce56fcd9a693d20a3877162dd0664f0b9529b2a560fd"} Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.133273 4871 generic.go:334] "Generic (PLEG): container finished" podID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerID="339d971690348c1a92f2ce56fcd9a693d20a3877162dd0664f0b9529b2a560fd" exitCode=0 Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.147060 4871 generic.go:334] "Generic (PLEG): container finished" podID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerID="4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5" exitCode=0 Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.147737 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbjpj" event={"ID":"cc78d2b6-52d6-413a-9f72-5b86a26ff79a","Type":"ContainerDied","Data":"4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5"} Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.154832 4871 generic.go:334] "Generic (PLEG): container finished" podID="91c74861-3e58-4603-b0fd-d791b0368823" containerID="32362cac514e2e27e1438c49e9e6a37586be3c32bb961e5261eab96d18d84a7b" exitCode=0 Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.154904 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjghv" event={"ID":"91c74861-3e58-4603-b0fd-d791b0368823","Type":"ContainerDied","Data":"32362cac514e2e27e1438c49e9e6a37586be3c32bb961e5261eab96d18d84a7b"} Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.156160 4871 generic.go:334] "Generic (PLEG): container finished" podID="9d97e32b-02b4-4435-bd25-81174b3e0da9" containerID="69605fa35c1363bbb7197afd08ab593a01b606b05174b53b3f359e14c60b2cd3" exitCode=0 Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.156249 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" event={"ID":"9d97e32b-02b4-4435-bd25-81174b3e0da9","Type":"ContainerDied","Data":"69605fa35c1363bbb7197afd08ab593a01b606b05174b53b3f359e14c60b2cd3"} Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.207877 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmnvw\" (UniqueName: \"kubernetes.io/projected/464efd4a-2401-457e-bf72-41ce7b9f2cac-kube-api-access-lmnvw\") pod \"marketplace-operator-79b997595-kgzf8\" (UID: \"464efd4a-2401-457e-bf72-41ce7b9f2cac\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.207951 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/464efd4a-2401-457e-bf72-41ce7b9f2cac-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kgzf8\" (UID: \"464efd4a-2401-457e-bf72-41ce7b9f2cac\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.207998 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/464efd4a-2401-457e-bf72-41ce7b9f2cac-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kgzf8\" (UID: \"464efd4a-2401-457e-bf72-41ce7b9f2cac\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.210184 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/464efd4a-2401-457e-bf72-41ce7b9f2cac-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kgzf8\" (UID: \"464efd4a-2401-457e-bf72-41ce7b9f2cac\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.216682 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/464efd4a-2401-457e-bf72-41ce7b9f2cac-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kgzf8\" (UID: \"464efd4a-2401-457e-bf72-41ce7b9f2cac\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.226274 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmnvw\" (UniqueName: \"kubernetes.io/projected/464efd4a-2401-457e-bf72-41ce7b9f2cac-kube-api-access-lmnvw\") pod \"marketplace-operator-79b997595-kgzf8\" (UID: \"464efd4a-2401-457e-bf72-41ce7b9f2cac\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.370393 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.376296 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.379823 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.388417 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.440052 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.442422 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.510634 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-trusted-ca\") pod \"9d97e32b-02b4-4435-bd25-81174b3e0da9\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.510686 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvqcw\" (UniqueName: \"kubernetes.io/projected/d868b268-45e4-480f-b84c-b74b51127ae6-kube-api-access-qvqcw\") pod \"d868b268-45e4-480f-b84c-b74b51127ae6\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.510728 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-utilities\") pod \"d868b268-45e4-480f-b84c-b74b51127ae6\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.510808 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-operator-metrics\") pod \"9d97e32b-02b4-4435-bd25-81174b3e0da9\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.510836 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-catalog-content\") pod \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.510864 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-utilities\") pod \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.510894 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cl8z\" (UniqueName: \"kubernetes.io/projected/9d97e32b-02b4-4435-bd25-81174b3e0da9-kube-api-access-8cl8z\") pod \"9d97e32b-02b4-4435-bd25-81174b3e0da9\" (UID: \"9d97e32b-02b4-4435-bd25-81174b3e0da9\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.510920 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5j22\" (UniqueName: \"kubernetes.io/projected/d02afe1f-19f0-448f-a9a1-0e7f19b39949-kube-api-access-x5j22\") pod \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\" (UID: \"d02afe1f-19f0-448f-a9a1-0e7f19b39949\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.510966 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-catalog-content\") pod \"d868b268-45e4-480f-b84c-b74b51127ae6\" (UID: \"d868b268-45e4-480f-b84c-b74b51127ae6\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.515754 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "9d97e32b-02b4-4435-bd25-81174b3e0da9" (UID: "9d97e32b-02b4-4435-bd25-81174b3e0da9"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.520528 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-utilities" (OuterVolumeSpecName: "utilities") pod "d02afe1f-19f0-448f-a9a1-0e7f19b39949" (UID: "d02afe1f-19f0-448f-a9a1-0e7f19b39949"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.520702 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d02afe1f-19f0-448f-a9a1-0e7f19b39949-kube-api-access-x5j22" (OuterVolumeSpecName: "kube-api-access-x5j22") pod "d02afe1f-19f0-448f-a9a1-0e7f19b39949" (UID: "d02afe1f-19f0-448f-a9a1-0e7f19b39949"). InnerVolumeSpecName "kube-api-access-x5j22". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.520715 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-utilities" (OuterVolumeSpecName: "utilities") pod "d868b268-45e4-480f-b84c-b74b51127ae6" (UID: "d868b268-45e4-480f-b84c-b74b51127ae6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.520847 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d868b268-45e4-480f-b84c-b74b51127ae6-kube-api-access-qvqcw" (OuterVolumeSpecName: "kube-api-access-qvqcw") pod "d868b268-45e4-480f-b84c-b74b51127ae6" (UID: "d868b268-45e4-480f-b84c-b74b51127ae6"). InnerVolumeSpecName "kube-api-access-qvqcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.521022 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "9d97e32b-02b4-4435-bd25-81174b3e0da9" (UID: "9d97e32b-02b4-4435-bd25-81174b3e0da9"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.526328 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d97e32b-02b4-4435-bd25-81174b3e0da9-kube-api-access-8cl8z" (OuterVolumeSpecName: "kube-api-access-8cl8z") pod "9d97e32b-02b4-4435-bd25-81174b3e0da9" (UID: "9d97e32b-02b4-4435-bd25-81174b3e0da9"). InnerVolumeSpecName "kube-api-access-8cl8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.580316 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d02afe1f-19f0-448f-a9a1-0e7f19b39949" (UID: "d02afe1f-19f0-448f-a9a1-0e7f19b39949"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.586391 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d868b268-45e4-480f-b84c-b74b51127ae6" (UID: "d868b268-45e4-480f-b84c-b74b51127ae6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.614243 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-catalog-content\") pod \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.614318 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-utilities\") pod \"91c74861-3e58-4603-b0fd-d791b0368823\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.616258 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9szcc\" (UniqueName: \"kubernetes.io/projected/91c74861-3e58-4603-b0fd-d791b0368823-kube-api-access-9szcc\") pod \"91c74861-3e58-4603-b0fd-d791b0368823\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.616296 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp6kt\" (UniqueName: \"kubernetes.io/projected/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-kube-api-access-dp6kt\") pod \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.616401 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-catalog-content\") pod \"91c74861-3e58-4603-b0fd-d791b0368823\" (UID: \"91c74861-3e58-4603-b0fd-d791b0368823\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.616895 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-utilities\") pod \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\" (UID: \"cc78d2b6-52d6-413a-9f72-5b86a26ff79a\") " Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617310 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617329 4871 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617342 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvqcw\" (UniqueName: \"kubernetes.io/projected/d868b268-45e4-480f-b84c-b74b51127ae6-kube-api-access-qvqcw\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617355 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d868b268-45e4-480f-b84c-b74b51127ae6-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617364 4871 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d97e32b-02b4-4435-bd25-81174b3e0da9-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617375 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617432 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d02afe1f-19f0-448f-a9a1-0e7f19b39949-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617444 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cl8z\" (UniqueName: \"kubernetes.io/projected/9d97e32b-02b4-4435-bd25-81174b3e0da9-kube-api-access-8cl8z\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617454 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5j22\" (UniqueName: \"kubernetes.io/projected/d02afe1f-19f0-448f-a9a1-0e7f19b39949-kube-api-access-x5j22\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.617662 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-utilities" (OuterVolumeSpecName: "utilities") pod "cc78d2b6-52d6-413a-9f72-5b86a26ff79a" (UID: "cc78d2b6-52d6-413a-9f72-5b86a26ff79a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.618839 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-utilities" (OuterVolumeSpecName: "utilities") pod "91c74861-3e58-4603-b0fd-d791b0368823" (UID: "91c74861-3e58-4603-b0fd-d791b0368823"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.619602 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91c74861-3e58-4603-b0fd-d791b0368823-kube-api-access-9szcc" (OuterVolumeSpecName: "kube-api-access-9szcc") pod "91c74861-3e58-4603-b0fd-d791b0368823" (UID: "91c74861-3e58-4603-b0fd-d791b0368823"). InnerVolumeSpecName "kube-api-access-9szcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.620409 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-kube-api-access-dp6kt" (OuterVolumeSpecName: "kube-api-access-dp6kt") pod "cc78d2b6-52d6-413a-9f72-5b86a26ff79a" (UID: "cc78d2b6-52d6-413a-9f72-5b86a26ff79a"). InnerVolumeSpecName "kube-api-access-dp6kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.626851 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kgzf8"] Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.637876 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc78d2b6-52d6-413a-9f72-5b86a26ff79a" (UID: "cc78d2b6-52d6-413a-9f72-5b86a26ff79a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.713241 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91c74861-3e58-4603-b0fd-d791b0368823" (UID: "91c74861-3e58-4603-b0fd-d791b0368823"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.719071 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.719195 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.719210 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.719221 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c74861-3e58-4603-b0fd-d791b0368823-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.719233 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9szcc\" (UniqueName: \"kubernetes.io/projected/91c74861-3e58-4603-b0fd-d791b0368823-kube-api-access-9szcc\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:40 crc kubenswrapper[4871]: I1007 22:14:40.719248 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp6kt\" (UniqueName: \"kubernetes.io/projected/cc78d2b6-52d6-413a-9f72-5b86a26ff79a-kube-api-access-dp6kt\") on node \"crc\" DevicePath \"\"" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.165116 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbjpj" event={"ID":"cc78d2b6-52d6-413a-9f72-5b86a26ff79a","Type":"ContainerDied","Data":"24e785dd251e7d481f5ec326b13c3d7299477b45789f53e55646923ff1457e90"} Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.165210 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbjpj" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.167027 4871 scope.go:117] "RemoveContainer" containerID="4eb8b5d88d3a9df3dafd064863bf913a37f66f42dbcfa576a08083f8816e37f5" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.168938 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwkd7" event={"ID":"d02afe1f-19f0-448f-a9a1-0e7f19b39949","Type":"ContainerDied","Data":"86f45828fcdfdc87168bf8dd3968cad59215b19e2aa9d96d32dea167aa7902dd"} Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.168986 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rwkd7" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.172099 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjghv" event={"ID":"91c74861-3e58-4603-b0fd-d791b0368823","Type":"ContainerDied","Data":"b6c217c4bb047a1c3e9394b338af1e6b6d9366865df84bbbbee1c66d0bfacf98"} Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.172179 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gjghv" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.176098 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" event={"ID":"464efd4a-2401-457e-bf72-41ce7b9f2cac","Type":"ContainerStarted","Data":"26f6b4562493bc1508e23e8880eddc5b23e63b72074898d11698765a425c075e"} Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.176137 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" event={"ID":"464efd4a-2401-457e-bf72-41ce7b9f2cac","Type":"ContainerStarted","Data":"d21a0ab107b743a9ed338d59d3f60cdd4062db7d575cf9b09b5959f18ab8c5c3"} Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.177894 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.177906 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kvwx2" event={"ID":"9d97e32b-02b4-4435-bd25-81174b3e0da9","Type":"ContainerDied","Data":"43fa43e7834ba9c6349858e1a3d5ca4c6fdfa5e1343b1eea6b2ddec55b325730"} Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.181004 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqpjz" event={"ID":"d868b268-45e4-480f-b84c-b74b51127ae6","Type":"ContainerDied","Data":"6f0aa10e85629533fc74aca60bc8b98e1552283c0fe10fc10c5fe66de85ec3ab"} Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.181061 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqpjz" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.185521 4871 scope.go:117] "RemoveContainer" containerID="43867ff257f30b4e76103c0af651971edc0c66a7290985354bddf8252e2dc22c" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.201285 4871 scope.go:117] "RemoveContainer" containerID="7f70dbc8a81ee237e3e4c88ea227ab4a60f12943ea0066a3b1a0644e941e9cc2" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.205324 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gjghv"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.214862 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gjghv"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.227186 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbjpj"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.236227 4871 scope.go:117] "RemoveContainer" containerID="339d971690348c1a92f2ce56fcd9a693d20a3877162dd0664f0b9529b2a560fd" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.239961 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbjpj"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.243088 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dqpjz"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.245471 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dqpjz"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.247937 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kvwx2"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.252660 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kvwx2"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.257173 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rwkd7"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.260541 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rwkd7"] Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.261176 4871 scope.go:117] "RemoveContainer" containerID="68858ef6583dfb1f51ded1c27ea80c0331fdb62fc1e2ae92b0d7d3dce359f076" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.271186 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" podStartSLOduration=2.271163756 podStartE2EDuration="2.271163756s" podCreationTimestamp="2025-10-07 22:14:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:14:41.268289536 +0000 UTC m=+355.070987609" watchObservedRunningTime="2025-10-07 22:14:41.271163756 +0000 UTC m=+355.073861829" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.278924 4871 scope.go:117] "RemoveContainer" containerID="b52292ec72c1b7b8bded703a8c25d859c5832accdf791a978969b78be1cac7fb" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.297202 4871 scope.go:117] "RemoveContainer" containerID="32362cac514e2e27e1438c49e9e6a37586be3c32bb961e5261eab96d18d84a7b" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.315985 4871 scope.go:117] "RemoveContainer" containerID="fc429956e4c0d480fb3c46005fb0cfedaf5839d46871c246f70ee338e6baf62d" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.340132 4871 scope.go:117] "RemoveContainer" containerID="2fe8798016613c1af5ec77168244b8079b68dc7b6781bc39d087a9ec47fb19cb" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.354737 4871 scope.go:117] "RemoveContainer" containerID="69605fa35c1363bbb7197afd08ab593a01b606b05174b53b3f359e14c60b2cd3" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.368969 4871 scope.go:117] "RemoveContainer" containerID="d7cca54b452a1f0159c8409f59e9da4cf5149f67361e1d0c8ccffad6de9adf27" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.380869 4871 scope.go:117] "RemoveContainer" containerID="1d1fb46f483f549bdaec907aa5f675dfe36ec10c80a9c2041f6f83a9e958fad8" Oct 07 22:14:41 crc kubenswrapper[4871]: I1007 22:14:41.399995 4871 scope.go:117] "RemoveContainer" containerID="11921cd710209bea5354890b3de2e53cc9330fd01a36f84d1b0377d5c5a64b81" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.083759 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4jd9w"] Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084394 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084410 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084423 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c74861-3e58-4603-b0fd-d791b0368823" containerName="extract-content" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084430 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c74861-3e58-4603-b0fd-d791b0368823" containerName="extract-content" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084440 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" containerName="extract-utilities" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084448 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" containerName="extract-utilities" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084461 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c74861-3e58-4603-b0fd-d791b0368823" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084467 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c74861-3e58-4603-b0fd-d791b0368823" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084477 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerName="extract-content" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084483 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerName="extract-content" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084489 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" containerName="extract-content" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084495 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" containerName="extract-content" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084502 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerName="extract-content" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084507 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerName="extract-content" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084516 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerName="extract-utilities" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084522 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerName="extract-utilities" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084531 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084537 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084547 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerName="extract-utilities" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084553 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerName="extract-utilities" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084562 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084569 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084576 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d97e32b-02b4-4435-bd25-81174b3e0da9" containerName="marketplace-operator" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084582 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d97e32b-02b4-4435-bd25-81174b3e0da9" containerName="marketplace-operator" Oct 07 22:14:42 crc kubenswrapper[4871]: E1007 22:14:42.084592 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c74861-3e58-4603-b0fd-d791b0368823" containerName="extract-utilities" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084597 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c74861-3e58-4603-b0fd-d791b0368823" containerName="extract-utilities" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084682 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084698 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084704 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d97e32b-02b4-4435-bd25-81174b3e0da9" containerName="marketplace-operator" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084712 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="91c74861-3e58-4603-b0fd-d791b0368823" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.084718 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" containerName="registry-server" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.085479 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.087815 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.103381 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4jd9w"] Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.194648 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.198597 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kgzf8" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.247152 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-catalog-content\") pod \"redhat-marketplace-4jd9w\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.247614 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-utilities\") pod \"redhat-marketplace-4jd9w\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.247752 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f76dw\" (UniqueName: \"kubernetes.io/projected/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-kube-api-access-f76dw\") pod \"redhat-marketplace-4jd9w\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.282759 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tsvn2"] Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.284106 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.286468 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.300333 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tsvn2"] Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.348879 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-catalog-content\") pod \"redhat-marketplace-4jd9w\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.348962 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-utilities\") pod \"redhat-marketplace-4jd9w\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.349016 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f76dw\" (UniqueName: \"kubernetes.io/projected/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-kube-api-access-f76dw\") pod \"redhat-marketplace-4jd9w\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.349653 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-catalog-content\") pod \"redhat-marketplace-4jd9w\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.349696 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-utilities\") pod \"redhat-marketplace-4jd9w\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.372399 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f76dw\" (UniqueName: \"kubernetes.io/projected/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-kube-api-access-f76dw\") pod \"redhat-marketplace-4jd9w\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.407369 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.450561 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw97v\" (UniqueName: \"kubernetes.io/projected/cef3b48f-0bfa-43f7-80ce-e58b5353637a-kube-api-access-zw97v\") pod \"redhat-operators-tsvn2\" (UID: \"cef3b48f-0bfa-43f7-80ce-e58b5353637a\") " pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.450950 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef3b48f-0bfa-43f7-80ce-e58b5353637a-utilities\") pod \"redhat-operators-tsvn2\" (UID: \"cef3b48f-0bfa-43f7-80ce-e58b5353637a\") " pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.451120 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef3b48f-0bfa-43f7-80ce-e58b5353637a-catalog-content\") pod \"redhat-operators-tsvn2\" (UID: \"cef3b48f-0bfa-43f7-80ce-e58b5353637a\") " pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.553950 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw97v\" (UniqueName: \"kubernetes.io/projected/cef3b48f-0bfa-43f7-80ce-e58b5353637a-kube-api-access-zw97v\") pod \"redhat-operators-tsvn2\" (UID: \"cef3b48f-0bfa-43f7-80ce-e58b5353637a\") " pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.554418 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef3b48f-0bfa-43f7-80ce-e58b5353637a-utilities\") pod \"redhat-operators-tsvn2\" (UID: \"cef3b48f-0bfa-43f7-80ce-e58b5353637a\") " pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.554459 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef3b48f-0bfa-43f7-80ce-e58b5353637a-catalog-content\") pod \"redhat-operators-tsvn2\" (UID: \"cef3b48f-0bfa-43f7-80ce-e58b5353637a\") " pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.554974 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef3b48f-0bfa-43f7-80ce-e58b5353637a-catalog-content\") pod \"redhat-operators-tsvn2\" (UID: \"cef3b48f-0bfa-43f7-80ce-e58b5353637a\") " pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.555441 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef3b48f-0bfa-43f7-80ce-e58b5353637a-utilities\") pod \"redhat-operators-tsvn2\" (UID: \"cef3b48f-0bfa-43f7-80ce-e58b5353637a\") " pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.572380 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw97v\" (UniqueName: \"kubernetes.io/projected/cef3b48f-0bfa-43f7-80ce-e58b5353637a-kube-api-access-zw97v\") pod \"redhat-operators-tsvn2\" (UID: \"cef3b48f-0bfa-43f7-80ce-e58b5353637a\") " pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.601025 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.608569 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4jd9w"] Oct 07 22:14:42 crc kubenswrapper[4871]: W1007 22:14:42.619152 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c3eefd2_0d10_4f32_86e5_f6d7af4a2bf2.slice/crio-339cbec5f04200f415d95f2c07a6f30e687fb298e20aac3961d71206f2f1141f WatchSource:0}: Error finding container 339cbec5f04200f415d95f2c07a6f30e687fb298e20aac3961d71206f2f1141f: Status 404 returned error can't find the container with id 339cbec5f04200f415d95f2c07a6f30e687fb298e20aac3961d71206f2f1141f Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.799673 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tsvn2"] Oct 07 22:14:42 crc kubenswrapper[4871]: W1007 22:14:42.836956 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcef3b48f_0bfa_43f7_80ce_e58b5353637a.slice/crio-28f6340832ac1b30d186ce45d4cf487791e00fce2c4f8add2449ae74997719d0 WatchSource:0}: Error finding container 28f6340832ac1b30d186ce45d4cf487791e00fce2c4f8add2449ae74997719d0: Status 404 returned error can't find the container with id 28f6340832ac1b30d186ce45d4cf487791e00fce2c4f8add2449ae74997719d0 Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.994317 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91c74861-3e58-4603-b0fd-d791b0368823" path="/var/lib/kubelet/pods/91c74861-3e58-4603-b0fd-d791b0368823/volumes" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.996480 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d97e32b-02b4-4435-bd25-81174b3e0da9" path="/var/lib/kubelet/pods/9d97e32b-02b4-4435-bd25-81174b3e0da9/volumes" Oct 07 22:14:42 crc kubenswrapper[4871]: I1007 22:14:42.997474 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc78d2b6-52d6-413a-9f72-5b86a26ff79a" path="/var/lib/kubelet/pods/cc78d2b6-52d6-413a-9f72-5b86a26ff79a/volumes" Oct 07 22:14:43 crc kubenswrapper[4871]: I1007 22:14:42.999974 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d02afe1f-19f0-448f-a9a1-0e7f19b39949" path="/var/lib/kubelet/pods/d02afe1f-19f0-448f-a9a1-0e7f19b39949/volumes" Oct 07 22:14:43 crc kubenswrapper[4871]: I1007 22:14:43.001986 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d868b268-45e4-480f-b84c-b74b51127ae6" path="/var/lib/kubelet/pods/d868b268-45e4-480f-b84c-b74b51127ae6/volumes" Oct 07 22:14:43 crc kubenswrapper[4871]: I1007 22:14:43.202751 4871 generic.go:334] "Generic (PLEG): container finished" podID="cef3b48f-0bfa-43f7-80ce-e58b5353637a" containerID="01b851e72fae1ba299f493bd06856762977b56b02c2ae7e6a09dbea638843f0f" exitCode=0 Oct 07 22:14:43 crc kubenswrapper[4871]: I1007 22:14:43.202861 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsvn2" event={"ID":"cef3b48f-0bfa-43f7-80ce-e58b5353637a","Type":"ContainerDied","Data":"01b851e72fae1ba299f493bd06856762977b56b02c2ae7e6a09dbea638843f0f"} Oct 07 22:14:43 crc kubenswrapper[4871]: I1007 22:14:43.202906 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsvn2" event={"ID":"cef3b48f-0bfa-43f7-80ce-e58b5353637a","Type":"ContainerStarted","Data":"28f6340832ac1b30d186ce45d4cf487791e00fce2c4f8add2449ae74997719d0"} Oct 07 22:14:43 crc kubenswrapper[4871]: I1007 22:14:43.208302 4871 generic.go:334] "Generic (PLEG): container finished" podID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerID="050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775" exitCode=0 Oct 07 22:14:43 crc kubenswrapper[4871]: I1007 22:14:43.208415 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4jd9w" event={"ID":"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2","Type":"ContainerDied","Data":"050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775"} Oct 07 22:14:43 crc kubenswrapper[4871]: I1007 22:14:43.208502 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4jd9w" event={"ID":"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2","Type":"ContainerStarted","Data":"339cbec5f04200f415d95f2c07a6f30e687fb298e20aac3961d71206f2f1141f"} Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.220186 4871 generic.go:334] "Generic (PLEG): container finished" podID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerID="7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f" exitCode=0 Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.220286 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4jd9w" event={"ID":"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2","Type":"ContainerDied","Data":"7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f"} Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.224040 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsvn2" event={"ID":"cef3b48f-0bfa-43f7-80ce-e58b5353637a","Type":"ContainerStarted","Data":"8658314de934ec1707b3752e18c2c5a52658f7b1e607683882c47d929e385466"} Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.484884 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5gn7g"] Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.487725 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.492203 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.500999 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5gn7g"] Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.501581 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-catalog-content\") pod \"certified-operators-5gn7g\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.501746 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lplmt\" (UniqueName: \"kubernetes.io/projected/ac1066e8-f230-4d5f-9a0c-639f2874fab5-kube-api-access-lplmt\") pod \"certified-operators-5gn7g\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.501918 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-utilities\") pod \"certified-operators-5gn7g\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.609090 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-catalog-content\") pod \"certified-operators-5gn7g\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.609561 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lplmt\" (UniqueName: \"kubernetes.io/projected/ac1066e8-f230-4d5f-9a0c-639f2874fab5-kube-api-access-lplmt\") pod \"certified-operators-5gn7g\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.611043 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-utilities\") pod \"certified-operators-5gn7g\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.611419 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-catalog-content\") pod \"certified-operators-5gn7g\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.611953 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-utilities\") pod \"certified-operators-5gn7g\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.648492 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lplmt\" (UniqueName: \"kubernetes.io/projected/ac1066e8-f230-4d5f-9a0c-639f2874fab5-kube-api-access-lplmt\") pod \"certified-operators-5gn7g\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.685221 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pn6qm"] Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.686932 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.692558 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.702743 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pn6qm"] Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.713073 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48jgg\" (UniqueName: \"kubernetes.io/projected/e67ed465-27ab-4f9b-a8ab-faf4b17f7437-kube-api-access-48jgg\") pod \"community-operators-pn6qm\" (UID: \"e67ed465-27ab-4f9b-a8ab-faf4b17f7437\") " pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.713168 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e67ed465-27ab-4f9b-a8ab-faf4b17f7437-utilities\") pod \"community-operators-pn6qm\" (UID: \"e67ed465-27ab-4f9b-a8ab-faf4b17f7437\") " pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.713197 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e67ed465-27ab-4f9b-a8ab-faf4b17f7437-catalog-content\") pod \"community-operators-pn6qm\" (UID: \"e67ed465-27ab-4f9b-a8ab-faf4b17f7437\") " pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.804436 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.814224 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48jgg\" (UniqueName: \"kubernetes.io/projected/e67ed465-27ab-4f9b-a8ab-faf4b17f7437-kube-api-access-48jgg\") pod \"community-operators-pn6qm\" (UID: \"e67ed465-27ab-4f9b-a8ab-faf4b17f7437\") " pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.814419 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e67ed465-27ab-4f9b-a8ab-faf4b17f7437-utilities\") pod \"community-operators-pn6qm\" (UID: \"e67ed465-27ab-4f9b-a8ab-faf4b17f7437\") " pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.814491 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e67ed465-27ab-4f9b-a8ab-faf4b17f7437-catalog-content\") pod \"community-operators-pn6qm\" (UID: \"e67ed465-27ab-4f9b-a8ab-faf4b17f7437\") " pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.815676 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e67ed465-27ab-4f9b-a8ab-faf4b17f7437-utilities\") pod \"community-operators-pn6qm\" (UID: \"e67ed465-27ab-4f9b-a8ab-faf4b17f7437\") " pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.816176 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e67ed465-27ab-4f9b-a8ab-faf4b17f7437-catalog-content\") pod \"community-operators-pn6qm\" (UID: \"e67ed465-27ab-4f9b-a8ab-faf4b17f7437\") " pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:44 crc kubenswrapper[4871]: I1007 22:14:44.832775 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48jgg\" (UniqueName: \"kubernetes.io/projected/e67ed465-27ab-4f9b-a8ab-faf4b17f7437-kube-api-access-48jgg\") pod \"community-operators-pn6qm\" (UID: \"e67ed465-27ab-4f9b-a8ab-faf4b17f7437\") " pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:45 crc kubenswrapper[4871]: I1007 22:14:45.026724 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5gn7g"] Oct 07 22:14:45 crc kubenswrapper[4871]: W1007 22:14:45.042183 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac1066e8_f230_4d5f_9a0c_639f2874fab5.slice/crio-30d69b02629c93d1b052a960dc03ddea92c245f58f220fc74145ac80aa7a876b WatchSource:0}: Error finding container 30d69b02629c93d1b052a960dc03ddea92c245f58f220fc74145ac80aa7a876b: Status 404 returned error can't find the container with id 30d69b02629c93d1b052a960dc03ddea92c245f58f220fc74145ac80aa7a876b Oct 07 22:14:45 crc kubenswrapper[4871]: I1007 22:14:45.090409 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:45 crc kubenswrapper[4871]: I1007 22:14:45.237724 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4jd9w" event={"ID":"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2","Type":"ContainerStarted","Data":"9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e"} Oct 07 22:14:45 crc kubenswrapper[4871]: I1007 22:14:45.268201 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gn7g" event={"ID":"ac1066e8-f230-4d5f-9a0c-639f2874fab5","Type":"ContainerStarted","Data":"cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03"} Oct 07 22:14:45 crc kubenswrapper[4871]: I1007 22:14:45.269495 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gn7g" event={"ID":"ac1066e8-f230-4d5f-9a0c-639f2874fab5","Type":"ContainerStarted","Data":"30d69b02629c93d1b052a960dc03ddea92c245f58f220fc74145ac80aa7a876b"} Oct 07 22:14:45 crc kubenswrapper[4871]: I1007 22:14:45.270239 4871 generic.go:334] "Generic (PLEG): container finished" podID="cef3b48f-0bfa-43f7-80ce-e58b5353637a" containerID="8658314de934ec1707b3752e18c2c5a52658f7b1e607683882c47d929e385466" exitCode=0 Oct 07 22:14:45 crc kubenswrapper[4871]: I1007 22:14:45.270306 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsvn2" event={"ID":"cef3b48f-0bfa-43f7-80ce-e58b5353637a","Type":"ContainerDied","Data":"8658314de934ec1707b3752e18c2c5a52658f7b1e607683882c47d929e385466"} Oct 07 22:14:45 crc kubenswrapper[4871]: I1007 22:14:45.295855 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4jd9w" podStartSLOduration=1.87949162 podStartE2EDuration="3.295824223s" podCreationTimestamp="2025-10-07 22:14:42 +0000 UTC" firstStartedPulling="2025-10-07 22:14:43.217148501 +0000 UTC m=+357.019846584" lastFinishedPulling="2025-10-07 22:14:44.633481114 +0000 UTC m=+358.436179187" observedRunningTime="2025-10-07 22:14:45.256539448 +0000 UTC m=+359.059237531" watchObservedRunningTime="2025-10-07 22:14:45.295824223 +0000 UTC m=+359.098522306" Oct 07 22:14:45 crc kubenswrapper[4871]: I1007 22:14:45.326402 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pn6qm"] Oct 07 22:14:45 crc kubenswrapper[4871]: W1007 22:14:45.340199 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode67ed465_27ab_4f9b_a8ab_faf4b17f7437.slice/crio-4a858f0cc185aa0595c367e4b64c6354ecd2c0c9fdfb9665516d97c8b40e0dac WatchSource:0}: Error finding container 4a858f0cc185aa0595c367e4b64c6354ecd2c0c9fdfb9665516d97c8b40e0dac: Status 404 returned error can't find the container with id 4a858f0cc185aa0595c367e4b64c6354ecd2c0c9fdfb9665516d97c8b40e0dac Oct 07 22:14:46 crc kubenswrapper[4871]: I1007 22:14:46.280499 4871 generic.go:334] "Generic (PLEG): container finished" podID="e67ed465-27ab-4f9b-a8ab-faf4b17f7437" containerID="1a885c81dfcc62931a8e86ef43551863a07cd607336d6b74aa696174bc10e06c" exitCode=0 Oct 07 22:14:46 crc kubenswrapper[4871]: I1007 22:14:46.281035 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn6qm" event={"ID":"e67ed465-27ab-4f9b-a8ab-faf4b17f7437","Type":"ContainerDied","Data":"1a885c81dfcc62931a8e86ef43551863a07cd607336d6b74aa696174bc10e06c"} Oct 07 22:14:46 crc kubenswrapper[4871]: I1007 22:14:46.281078 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn6qm" event={"ID":"e67ed465-27ab-4f9b-a8ab-faf4b17f7437","Type":"ContainerStarted","Data":"4a858f0cc185aa0595c367e4b64c6354ecd2c0c9fdfb9665516d97c8b40e0dac"} Oct 07 22:14:46 crc kubenswrapper[4871]: I1007 22:14:46.288301 4871 generic.go:334] "Generic (PLEG): container finished" podID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerID="cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03" exitCode=0 Oct 07 22:14:46 crc kubenswrapper[4871]: I1007 22:14:46.288878 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gn7g" event={"ID":"ac1066e8-f230-4d5f-9a0c-639f2874fab5","Type":"ContainerDied","Data":"cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03"} Oct 07 22:14:46 crc kubenswrapper[4871]: I1007 22:14:46.288935 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gn7g" event={"ID":"ac1066e8-f230-4d5f-9a0c-639f2874fab5","Type":"ContainerStarted","Data":"43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28"} Oct 07 22:14:46 crc kubenswrapper[4871]: I1007 22:14:46.293372 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsvn2" event={"ID":"cef3b48f-0bfa-43f7-80ce-e58b5353637a","Type":"ContainerStarted","Data":"cfb1d68ee6d4c78d4fd4befdc7a62fe0ae37c62c71c7e9f7b4070abc6383cb40"} Oct 07 22:14:47 crc kubenswrapper[4871]: I1007 22:14:47.299062 4871 generic.go:334] "Generic (PLEG): container finished" podID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerID="43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28" exitCode=0 Oct 07 22:14:47 crc kubenswrapper[4871]: I1007 22:14:47.299266 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gn7g" event={"ID":"ac1066e8-f230-4d5f-9a0c-639f2874fab5","Type":"ContainerDied","Data":"43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28"} Oct 07 22:14:47 crc kubenswrapper[4871]: I1007 22:14:47.301392 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn6qm" event={"ID":"e67ed465-27ab-4f9b-a8ab-faf4b17f7437","Type":"ContainerStarted","Data":"8af90ec33c2135104895ab992f361e68a7d392b4295b433e96cd7557aba2688c"} Oct 07 22:14:47 crc kubenswrapper[4871]: I1007 22:14:47.319943 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tsvn2" podStartSLOduration=2.775525698 podStartE2EDuration="5.319919638s" podCreationTimestamp="2025-10-07 22:14:42 +0000 UTC" firstStartedPulling="2025-10-07 22:14:43.205063077 +0000 UTC m=+357.007761150" lastFinishedPulling="2025-10-07 22:14:45.749457017 +0000 UTC m=+359.552155090" observedRunningTime="2025-10-07 22:14:46.341127934 +0000 UTC m=+360.143826017" watchObservedRunningTime="2025-10-07 22:14:47.319919638 +0000 UTC m=+361.122617711" Oct 07 22:14:48 crc kubenswrapper[4871]: I1007 22:14:48.309134 4871 generic.go:334] "Generic (PLEG): container finished" podID="e67ed465-27ab-4f9b-a8ab-faf4b17f7437" containerID="8af90ec33c2135104895ab992f361e68a7d392b4295b433e96cd7557aba2688c" exitCode=0 Oct 07 22:14:48 crc kubenswrapper[4871]: I1007 22:14:48.310075 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn6qm" event={"ID":"e67ed465-27ab-4f9b-a8ab-faf4b17f7437","Type":"ContainerDied","Data":"8af90ec33c2135104895ab992f361e68a7d392b4295b433e96cd7557aba2688c"} Oct 07 22:14:49 crc kubenswrapper[4871]: I1007 22:14:49.320236 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn6qm" event={"ID":"e67ed465-27ab-4f9b-a8ab-faf4b17f7437","Type":"ContainerStarted","Data":"27a5766e86dc07e2b78aae44371a3d17a33fe84aca46403891858de417a69343"} Oct 07 22:14:49 crc kubenswrapper[4871]: I1007 22:14:49.324294 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gn7g" event={"ID":"ac1066e8-f230-4d5f-9a0c-639f2874fab5","Type":"ContainerStarted","Data":"4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e"} Oct 07 22:14:49 crc kubenswrapper[4871]: I1007 22:14:49.339127 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pn6qm" podStartSLOduration=2.704938186 podStartE2EDuration="5.339104216s" podCreationTimestamp="2025-10-07 22:14:44 +0000 UTC" firstStartedPulling="2025-10-07 22:14:46.283492752 +0000 UTC m=+360.086190825" lastFinishedPulling="2025-10-07 22:14:48.917658782 +0000 UTC m=+362.720356855" observedRunningTime="2025-10-07 22:14:49.337222254 +0000 UTC m=+363.139920337" watchObservedRunningTime="2025-10-07 22:14:49.339104216 +0000 UTC m=+363.141802299" Oct 07 22:14:49 crc kubenswrapper[4871]: I1007 22:14:49.358440 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5gn7g" podStartSLOduration=2.907693497 podStartE2EDuration="5.358413569s" podCreationTimestamp="2025-10-07 22:14:44 +0000 UTC" firstStartedPulling="2025-10-07 22:14:45.269554197 +0000 UTC m=+359.072252270" lastFinishedPulling="2025-10-07 22:14:47.720274269 +0000 UTC m=+361.522972342" observedRunningTime="2025-10-07 22:14:49.356753584 +0000 UTC m=+363.159451657" watchObservedRunningTime="2025-10-07 22:14:49.358413569 +0000 UTC m=+363.161111642" Oct 07 22:14:52 crc kubenswrapper[4871]: I1007 22:14:52.408699 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:52 crc kubenswrapper[4871]: I1007 22:14:52.411013 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:52 crc kubenswrapper[4871]: I1007 22:14:52.486422 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:52 crc kubenswrapper[4871]: I1007 22:14:52.601867 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:52 crc kubenswrapper[4871]: I1007 22:14:52.601920 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:52 crc kubenswrapper[4871]: I1007 22:14:52.653996 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:53 crc kubenswrapper[4871]: I1007 22:14:53.398813 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tsvn2" Oct 07 22:14:53 crc kubenswrapper[4871]: I1007 22:14:53.424664 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:14:54 crc kubenswrapper[4871]: I1007 22:14:54.805605 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:54 crc kubenswrapper[4871]: I1007 22:14:54.806249 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:54 crc kubenswrapper[4871]: I1007 22:14:54.852298 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:55 crc kubenswrapper[4871]: I1007 22:14:55.091027 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:55 crc kubenswrapper[4871]: I1007 22:14:55.091116 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:55 crc kubenswrapper[4871]: I1007 22:14:55.135278 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:14:55 crc kubenswrapper[4871]: I1007 22:14:55.403395 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 22:14:55 crc kubenswrapper[4871]: I1007 22:14:55.408632 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pn6qm" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.152618 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp"] Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.153724 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.156364 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.156477 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.171427 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp"] Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.256415 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd86p\" (UniqueName: \"kubernetes.io/projected/0e7542f8-e26a-48e6-ba07-655b8d63b458-kube-api-access-hd86p\") pod \"collect-profiles-29331255-cb9hp\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.258217 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e7542f8-e26a-48e6-ba07-655b8d63b458-config-volume\") pod \"collect-profiles-29331255-cb9hp\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.258289 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e7542f8-e26a-48e6-ba07-655b8d63b458-secret-volume\") pod \"collect-profiles-29331255-cb9hp\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.360465 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd86p\" (UniqueName: \"kubernetes.io/projected/0e7542f8-e26a-48e6-ba07-655b8d63b458-kube-api-access-hd86p\") pod \"collect-profiles-29331255-cb9hp\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.360554 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e7542f8-e26a-48e6-ba07-655b8d63b458-config-volume\") pod \"collect-profiles-29331255-cb9hp\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.360608 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e7542f8-e26a-48e6-ba07-655b8d63b458-secret-volume\") pod \"collect-profiles-29331255-cb9hp\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.362657 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e7542f8-e26a-48e6-ba07-655b8d63b458-config-volume\") pod \"collect-profiles-29331255-cb9hp\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.369125 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e7542f8-e26a-48e6-ba07-655b8d63b458-secret-volume\") pod \"collect-profiles-29331255-cb9hp\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.379734 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd86p\" (UniqueName: \"kubernetes.io/projected/0e7542f8-e26a-48e6-ba07-655b8d63b458-kube-api-access-hd86p\") pod \"collect-profiles-29331255-cb9hp\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.472521 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:00 crc kubenswrapper[4871]: I1007 22:15:00.919444 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp"] Oct 07 22:15:00 crc kubenswrapper[4871]: W1007 22:15:00.925985 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e7542f8_e26a_48e6_ba07_655b8d63b458.slice/crio-7d850a2efc4cf348762e45588e9c37e3568717ede4fdba13da56fa06985f7ec8 WatchSource:0}: Error finding container 7d850a2efc4cf348762e45588e9c37e3568717ede4fdba13da56fa06985f7ec8: Status 404 returned error can't find the container with id 7d850a2efc4cf348762e45588e9c37e3568717ede4fdba13da56fa06985f7ec8 Oct 07 22:15:01 crc kubenswrapper[4871]: I1007 22:15:01.409665 4871 generic.go:334] "Generic (PLEG): container finished" podID="0e7542f8-e26a-48e6-ba07-655b8d63b458" containerID="871e2f33d5b297daf81fb62d40b9653b0696c86746858e34f8908e0ea9d4d33d" exitCode=0 Oct 07 22:15:01 crc kubenswrapper[4871]: I1007 22:15:01.409729 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" event={"ID":"0e7542f8-e26a-48e6-ba07-655b8d63b458","Type":"ContainerDied","Data":"871e2f33d5b297daf81fb62d40b9653b0696c86746858e34f8908e0ea9d4d33d"} Oct 07 22:15:01 crc kubenswrapper[4871]: I1007 22:15:01.409766 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" event={"ID":"0e7542f8-e26a-48e6-ba07-655b8d63b458","Type":"ContainerStarted","Data":"7d850a2efc4cf348762e45588e9c37e3568717ede4fdba13da56fa06985f7ec8"} Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.704579 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.795751 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd86p\" (UniqueName: \"kubernetes.io/projected/0e7542f8-e26a-48e6-ba07-655b8d63b458-kube-api-access-hd86p\") pod \"0e7542f8-e26a-48e6-ba07-655b8d63b458\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.795844 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e7542f8-e26a-48e6-ba07-655b8d63b458-config-volume\") pod \"0e7542f8-e26a-48e6-ba07-655b8d63b458\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.795889 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e7542f8-e26a-48e6-ba07-655b8d63b458-secret-volume\") pod \"0e7542f8-e26a-48e6-ba07-655b8d63b458\" (UID: \"0e7542f8-e26a-48e6-ba07-655b8d63b458\") " Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.797034 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e7542f8-e26a-48e6-ba07-655b8d63b458-config-volume" (OuterVolumeSpecName: "config-volume") pod "0e7542f8-e26a-48e6-ba07-655b8d63b458" (UID: "0e7542f8-e26a-48e6-ba07-655b8d63b458"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.804214 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e7542f8-e26a-48e6-ba07-655b8d63b458-kube-api-access-hd86p" (OuterVolumeSpecName: "kube-api-access-hd86p") pod "0e7542f8-e26a-48e6-ba07-655b8d63b458" (UID: "0e7542f8-e26a-48e6-ba07-655b8d63b458"). InnerVolumeSpecName "kube-api-access-hd86p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.804731 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e7542f8-e26a-48e6-ba07-655b8d63b458-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0e7542f8-e26a-48e6-ba07-655b8d63b458" (UID: "0e7542f8-e26a-48e6-ba07-655b8d63b458"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.897606 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd86p\" (UniqueName: \"kubernetes.io/projected/0e7542f8-e26a-48e6-ba07-655b8d63b458-kube-api-access-hd86p\") on node \"crc\" DevicePath \"\"" Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.897674 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e7542f8-e26a-48e6-ba07-655b8d63b458-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:15:02 crc kubenswrapper[4871]: I1007 22:15:02.897696 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e7542f8-e26a-48e6-ba07-655b8d63b458-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:15:03 crc kubenswrapper[4871]: I1007 22:15:03.424078 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" event={"ID":"0e7542f8-e26a-48e6-ba07-655b8d63b458","Type":"ContainerDied","Data":"7d850a2efc4cf348762e45588e9c37e3568717ede4fdba13da56fa06985f7ec8"} Oct 07 22:15:03 crc kubenswrapper[4871]: I1007 22:15:03.424145 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d850a2efc4cf348762e45588e9c37e3568717ede4fdba13da56fa06985f7ec8" Oct 07 22:15:03 crc kubenswrapper[4871]: I1007 22:15:03.424164 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp" Oct 07 22:15:05 crc kubenswrapper[4871]: I1007 22:15:05.512594 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:15:05 crc kubenswrapper[4871]: I1007 22:15:05.513544 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:15:35 crc kubenswrapper[4871]: I1007 22:15:35.512473 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:15:35 crc kubenswrapper[4871]: I1007 22:15:35.513338 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:15:35 crc kubenswrapper[4871]: I1007 22:15:35.513416 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:15:35 crc kubenswrapper[4871]: I1007 22:15:35.514159 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9e9168b99e93de9001a9d600cd6b89e0f408725c87f4710f2d8b93ae3777f08a"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:15:35 crc kubenswrapper[4871]: I1007 22:15:35.514256 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://9e9168b99e93de9001a9d600cd6b89e0f408725c87f4710f2d8b93ae3777f08a" gracePeriod=600 Oct 07 22:15:35 crc kubenswrapper[4871]: I1007 22:15:35.675492 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="9e9168b99e93de9001a9d600cd6b89e0f408725c87f4710f2d8b93ae3777f08a" exitCode=0 Oct 07 22:15:35 crc kubenswrapper[4871]: I1007 22:15:35.675552 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"9e9168b99e93de9001a9d600cd6b89e0f408725c87f4710f2d8b93ae3777f08a"} Oct 07 22:15:35 crc kubenswrapper[4871]: I1007 22:15:35.675619 4871 scope.go:117] "RemoveContainer" containerID="5cee6f7924cf06c66c9db32a857f2680f9cbc546f32b7f10c1911d215ccbe002" Oct 07 22:15:36 crc kubenswrapper[4871]: I1007 22:15:36.685559 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"fe9062423d89d34c8005bfab202352c48e57c724729bdae8f94451cff70b4074"} Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.780091 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bdths"] Oct 07 22:16:11 crc kubenswrapper[4871]: E1007 22:16:11.783105 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e7542f8-e26a-48e6-ba07-655b8d63b458" containerName="collect-profiles" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.783226 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e7542f8-e26a-48e6-ba07-655b8d63b458" containerName="collect-profiles" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.783442 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e7542f8-e26a-48e6-ba07-655b8d63b458" containerName="collect-profiles" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.784077 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.791785 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bdths"] Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.884916 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.884973 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-registry-tls\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.885006 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj24s\" (UniqueName: \"kubernetes.io/projected/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-kube-api-access-xj24s\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.885028 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-bound-sa-token\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.885067 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.885090 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.885149 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-trusted-ca\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.885204 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-registry-certificates\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.933952 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.987142 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.987226 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-trusted-ca\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.987291 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-registry-certificates\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.987341 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.987373 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-registry-tls\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.987405 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj24s\" (UniqueName: \"kubernetes.io/projected/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-kube-api-access-xj24s\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.987430 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-bound-sa-token\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.988767 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.989306 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-registry-certificates\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.990112 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-trusted-ca\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.995482 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-registry-tls\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:11 crc kubenswrapper[4871]: I1007 22:16:11.996114 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:12 crc kubenswrapper[4871]: I1007 22:16:12.003983 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-bound-sa-token\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:12 crc kubenswrapper[4871]: I1007 22:16:12.004946 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj24s\" (UniqueName: \"kubernetes.io/projected/a12c52d2-9d5b-4d2b-93f7-a9224b5b7992-kube-api-access-xj24s\") pod \"image-registry-66df7c8f76-bdths\" (UID: \"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992\") " pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:12 crc kubenswrapper[4871]: I1007 22:16:12.141351 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:12 crc kubenswrapper[4871]: I1007 22:16:12.399571 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bdths"] Oct 07 22:16:12 crc kubenswrapper[4871]: I1007 22:16:12.950547 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bdths" event={"ID":"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992","Type":"ContainerStarted","Data":"de7502aead78b5a027797825b41468722340481a22700eca729724db205c203e"} Oct 07 22:16:12 crc kubenswrapper[4871]: I1007 22:16:12.950938 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bdths" event={"ID":"a12c52d2-9d5b-4d2b-93f7-a9224b5b7992","Type":"ContainerStarted","Data":"23b8174ad21dc4f2d253ae01ebee607f4c50a4d2996af8041557bb54289b8b7c"} Oct 07 22:16:12 crc kubenswrapper[4871]: I1007 22:16:12.951054 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:12 crc kubenswrapper[4871]: I1007 22:16:12.979948 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bdths" podStartSLOduration=1.979925653 podStartE2EDuration="1.979925653s" podCreationTimestamp="2025-10-07 22:16:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:16:12.978098603 +0000 UTC m=+446.780796676" watchObservedRunningTime="2025-10-07 22:16:12.979925653 +0000 UTC m=+446.782623726" Oct 07 22:16:32 crc kubenswrapper[4871]: I1007 22:16:32.146019 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bdths" Oct 07 22:16:32 crc kubenswrapper[4871]: I1007 22:16:32.211587 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rfkm7"] Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.253763 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" podUID="3689423d-c1f1-46c8-9ea5-165147a8a286" containerName="registry" containerID="cri-o://51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358" gracePeriod=30 Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.693888 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.838264 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-certificates\") pod \"3689423d-c1f1-46c8-9ea5-165147a8a286\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.838968 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-bound-sa-token\") pod \"3689423d-c1f1-46c8-9ea5-165147a8a286\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.839199 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3689423d-c1f1-46c8-9ea5-165147a8a286-installation-pull-secrets\") pod \"3689423d-c1f1-46c8-9ea5-165147a8a286\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.839492 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-tls\") pod \"3689423d-c1f1-46c8-9ea5-165147a8a286\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.839842 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3689423d-c1f1-46c8-9ea5-165147a8a286-ca-trust-extracted\") pod \"3689423d-c1f1-46c8-9ea5-165147a8a286\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.840239 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"3689423d-c1f1-46c8-9ea5-165147a8a286\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.840488 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-trusted-ca\") pod \"3689423d-c1f1-46c8-9ea5-165147a8a286\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.840718 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgrft\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-kube-api-access-rgrft\") pod \"3689423d-c1f1-46c8-9ea5-165147a8a286\" (UID: \"3689423d-c1f1-46c8-9ea5-165147a8a286\") " Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.842662 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "3689423d-c1f1-46c8-9ea5-165147a8a286" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.842781 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "3689423d-c1f1-46c8-9ea5-165147a8a286" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.848264 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3689423d-c1f1-46c8-9ea5-165147a8a286-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "3689423d-c1f1-46c8-9ea5-165147a8a286" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.849203 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "3689423d-c1f1-46c8-9ea5-165147a8a286" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.849418 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "3689423d-c1f1-46c8-9ea5-165147a8a286" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.849733 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-kube-api-access-rgrft" (OuterVolumeSpecName: "kube-api-access-rgrft") pod "3689423d-c1f1-46c8-9ea5-165147a8a286" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286"). InnerVolumeSpecName "kube-api-access-rgrft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.852918 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "3689423d-c1f1-46c8-9ea5-165147a8a286" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.874737 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3689423d-c1f1-46c8-9ea5-165147a8a286-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "3689423d-c1f1-46c8-9ea5-165147a8a286" (UID: "3689423d-c1f1-46c8-9ea5-165147a8a286"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.943031 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgrft\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-kube-api-access-rgrft\") on node \"crc\" DevicePath \"\"" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.943090 4871 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3689423d-c1f1-46c8-9ea5-165147a8a286-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.943108 4871 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.943130 4871 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.943149 4871 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3689423d-c1f1-46c8-9ea5-165147a8a286-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.943169 4871 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3689423d-c1f1-46c8-9ea5-165147a8a286-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 07 22:16:57 crc kubenswrapper[4871]: I1007 22:16:57.943185 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3689423d-c1f1-46c8-9ea5-165147a8a286-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.249585 4871 generic.go:334] "Generic (PLEG): container finished" podID="3689423d-c1f1-46c8-9ea5-165147a8a286" containerID="51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358" exitCode=0 Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.249651 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" event={"ID":"3689423d-c1f1-46c8-9ea5-165147a8a286","Type":"ContainerDied","Data":"51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358"} Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.249692 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.249718 4871 scope.go:117] "RemoveContainer" containerID="51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358" Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.249698 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" event={"ID":"3689423d-c1f1-46c8-9ea5-165147a8a286","Type":"ContainerDied","Data":"5b729518d75ab356a8cbda851f0ba071062df62fd0d10ef18e236e225ab99161"} Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.289669 4871 scope.go:117] "RemoveContainer" containerID="51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358" Oct 07 22:16:58 crc kubenswrapper[4871]: E1007 22:16:58.294370 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358\": container with ID starting with 51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358 not found: ID does not exist" containerID="51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358" Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.294490 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358"} err="failed to get container status \"51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358\": rpc error: code = NotFound desc = could not find container \"51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358\": container with ID starting with 51df8a578554094dd961c29caf500846658517ef9ff29a0d59a774e2ef66f358 not found: ID does not exist" Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.303788 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rfkm7"] Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.310304 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rfkm7"] Oct 07 22:16:58 crc kubenswrapper[4871]: I1007 22:16:58.995480 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3689423d-c1f1-46c8-9ea5-165147a8a286" path="/var/lib/kubelet/pods/3689423d-c1f1-46c8-9ea5-165147a8a286/volumes" Oct 07 22:17:02 crc kubenswrapper[4871]: I1007 22:17:02.501048 4871 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-rfkm7 container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.16:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 22:17:02 crc kubenswrapper[4871]: I1007 22:17:02.501689 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-rfkm7" podUID="3689423d-c1f1-46c8-9ea5-165147a8a286" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.16:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 07 22:17:35 crc kubenswrapper[4871]: I1007 22:17:35.512751 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:17:35 crc kubenswrapper[4871]: I1007 22:17:35.513818 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:18:05 crc kubenswrapper[4871]: I1007 22:18:05.512739 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:18:05 crc kubenswrapper[4871]: I1007 22:18:05.513694 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:18:35 crc kubenswrapper[4871]: I1007 22:18:35.513344 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:18:35 crc kubenswrapper[4871]: I1007 22:18:35.514369 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:18:35 crc kubenswrapper[4871]: I1007 22:18:35.514440 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:18:35 crc kubenswrapper[4871]: I1007 22:18:35.515190 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe9062423d89d34c8005bfab202352c48e57c724729bdae8f94451cff70b4074"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:18:35 crc kubenswrapper[4871]: I1007 22:18:35.515252 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://fe9062423d89d34c8005bfab202352c48e57c724729bdae8f94451cff70b4074" gracePeriod=600 Oct 07 22:18:35 crc kubenswrapper[4871]: I1007 22:18:35.925414 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="fe9062423d89d34c8005bfab202352c48e57c724729bdae8f94451cff70b4074" exitCode=0 Oct 07 22:18:35 crc kubenswrapper[4871]: I1007 22:18:35.925502 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"fe9062423d89d34c8005bfab202352c48e57c724729bdae8f94451cff70b4074"} Oct 07 22:18:35 crc kubenswrapper[4871]: I1007 22:18:35.925899 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"f191ac46661d036cac4f07f7277192def3c4b3a3a733b08403472441f085e8e0"} Oct 07 22:18:35 crc kubenswrapper[4871]: I1007 22:18:35.925961 4871 scope.go:117] "RemoveContainer" containerID="9e9168b99e93de9001a9d600cd6b89e0f408725c87f4710f2d8b93ae3777f08a" Oct 07 22:20:35 crc kubenswrapper[4871]: I1007 22:20:35.512663 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:20:35 crc kubenswrapper[4871]: I1007 22:20:35.514070 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:21:05 crc kubenswrapper[4871]: I1007 22:21:05.512390 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:21:05 crc kubenswrapper[4871]: I1007 22:21:05.513554 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.136335 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7lxmn"] Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.137444 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" podUID="0609d7b1-7443-4bcd-935e-ca2a61a359b7" containerName="controller-manager" containerID="cri-o://f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e" gracePeriod=30 Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.223657 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2"] Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.224637 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" podUID="03ce8147-0273-421d-9263-3d4eb06bfcfe" containerName="route-controller-manager" containerID="cri-o://0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796" gracePeriod=30 Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.510561 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.527584 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0609d7b1-7443-4bcd-935e-ca2a61a359b7-serving-cert\") pod \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.528152 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-client-ca\") pod \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.528202 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-config\") pod \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.528262 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbrc9\" (UniqueName: \"kubernetes.io/projected/0609d7b1-7443-4bcd-935e-ca2a61a359b7-kube-api-access-gbrc9\") pod \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.528381 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-proxy-ca-bundles\") pod \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\" (UID: \"0609d7b1-7443-4bcd-935e-ca2a61a359b7\") " Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.529543 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0609d7b1-7443-4bcd-935e-ca2a61a359b7" (UID: "0609d7b1-7443-4bcd-935e-ca2a61a359b7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.530162 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "0609d7b1-7443-4bcd-935e-ca2a61a359b7" (UID: "0609d7b1-7443-4bcd-935e-ca2a61a359b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.531556 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-config" (OuterVolumeSpecName: "config") pod "0609d7b1-7443-4bcd-935e-ca2a61a359b7" (UID: "0609d7b1-7443-4bcd-935e-ca2a61a359b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.542900 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0609d7b1-7443-4bcd-935e-ca2a61a359b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0609d7b1-7443-4bcd-935e-ca2a61a359b7" (UID: "0609d7b1-7443-4bcd-935e-ca2a61a359b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.553381 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0609d7b1-7443-4bcd-935e-ca2a61a359b7-kube-api-access-gbrc9" (OuterVolumeSpecName: "kube-api-access-gbrc9") pod "0609d7b1-7443-4bcd-935e-ca2a61a359b7" (UID: "0609d7b1-7443-4bcd-935e-ca2a61a359b7"). InnerVolumeSpecName "kube-api-access-gbrc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.595778 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.630001 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce8147-0273-421d-9263-3d4eb06bfcfe-serving-cert\") pod \"03ce8147-0273-421d-9263-3d4eb06bfcfe\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.630095 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-client-ca\") pod \"03ce8147-0273-421d-9263-3d4eb06bfcfe\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.630144 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-config\") pod \"03ce8147-0273-421d-9263-3d4eb06bfcfe\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.630307 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p2jk\" (UniqueName: \"kubernetes.io/projected/03ce8147-0273-421d-9263-3d4eb06bfcfe-kube-api-access-9p2jk\") pod \"03ce8147-0273-421d-9263-3d4eb06bfcfe\" (UID: \"03ce8147-0273-421d-9263-3d4eb06bfcfe\") " Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.631538 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-client-ca" (OuterVolumeSpecName: "client-ca") pod "03ce8147-0273-421d-9263-3d4eb06bfcfe" (UID: "03ce8147-0273-421d-9263-3d4eb06bfcfe"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.632076 4871 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.632095 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0609d7b1-7443-4bcd-935e-ca2a61a359b7-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.632104 4871 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.632114 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0609d7b1-7443-4bcd-935e-ca2a61a359b7-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.632124 4871 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.632138 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbrc9\" (UniqueName: \"kubernetes.io/projected/0609d7b1-7443-4bcd-935e-ca2a61a359b7-kube-api-access-gbrc9\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.633311 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-config" (OuterVolumeSpecName: "config") pod "03ce8147-0273-421d-9263-3d4eb06bfcfe" (UID: "03ce8147-0273-421d-9263-3d4eb06bfcfe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.635734 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ce8147-0273-421d-9263-3d4eb06bfcfe-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "03ce8147-0273-421d-9263-3d4eb06bfcfe" (UID: "03ce8147-0273-421d-9263-3d4eb06bfcfe"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.636059 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03ce8147-0273-421d-9263-3d4eb06bfcfe-kube-api-access-9p2jk" (OuterVolumeSpecName: "kube-api-access-9p2jk") pod "03ce8147-0273-421d-9263-3d4eb06bfcfe" (UID: "03ce8147-0273-421d-9263-3d4eb06bfcfe"). InnerVolumeSpecName "kube-api-access-9p2jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.733276 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p2jk\" (UniqueName: \"kubernetes.io/projected/03ce8147-0273-421d-9263-3d4eb06bfcfe-kube-api-access-9p2jk\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.733313 4871 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ce8147-0273-421d-9263-3d4eb06bfcfe-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:13 crc kubenswrapper[4871]: I1007 22:21:13.733328 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ce8147-0273-421d-9263-3d4eb06bfcfe-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.109499 4871 generic.go:334] "Generic (PLEG): container finished" podID="03ce8147-0273-421d-9263-3d4eb06bfcfe" containerID="0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796" exitCode=0 Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.109588 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" event={"ID":"03ce8147-0273-421d-9263-3d4eb06bfcfe","Type":"ContainerDied","Data":"0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796"} Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.109622 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.109756 4871 scope.go:117] "RemoveContainer" containerID="0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.109727 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2" event={"ID":"03ce8147-0273-421d-9263-3d4eb06bfcfe","Type":"ContainerDied","Data":"2213d8d2ac1334eb918233091e1b977693e60ddc8dfb859044fb09cbcb836666"} Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.112251 4871 generic.go:334] "Generic (PLEG): container finished" podID="0609d7b1-7443-4bcd-935e-ca2a61a359b7" containerID="f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e" exitCode=0 Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.112290 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" event={"ID":"0609d7b1-7443-4bcd-935e-ca2a61a359b7","Type":"ContainerDied","Data":"f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e"} Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.112328 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" event={"ID":"0609d7b1-7443-4bcd-935e-ca2a61a359b7","Type":"ContainerDied","Data":"fb3cecb1a77b5b1156ad94c9c416959ab89ab44f6cd9363393761286eb16efea"} Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.112402 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7lxmn" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.140576 4871 scope.go:117] "RemoveContainer" containerID="0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796" Oct 07 22:21:14 crc kubenswrapper[4871]: E1007 22:21:14.143366 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796\": container with ID starting with 0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796 not found: ID does not exist" containerID="0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.143475 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796"} err="failed to get container status \"0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796\": rpc error: code = NotFound desc = could not find container \"0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796\": container with ID starting with 0ca55f6c4b3164fa0d56a183e2b17fb5386bab5671ccd4e7d822b8cb2e44f796 not found: ID does not exist" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.143543 4871 scope.go:117] "RemoveContainer" containerID="f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.157627 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7lxmn"] Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.163475 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7lxmn"] Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.174016 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2"] Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.175657 4871 scope.go:117] "RemoveContainer" containerID="f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e" Oct 07 22:21:14 crc kubenswrapper[4871]: E1007 22:21:14.176433 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e\": container with ID starting with f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e not found: ID does not exist" containerID="f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.176501 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e"} err="failed to get container status \"f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e\": rpc error: code = NotFound desc = could not find container \"f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e\": container with ID starting with f5d2d60b68f33c19dbf5be909555aaaac1ba1fa37291acc060a5af2b26827e8e not found: ID does not exist" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.183664 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48wk2"] Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.808029 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf"] Oct 07 22:21:14 crc kubenswrapper[4871]: E1007 22:21:14.808935 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3689423d-c1f1-46c8-9ea5-165147a8a286" containerName="registry" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.808979 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3689423d-c1f1-46c8-9ea5-165147a8a286" containerName="registry" Oct 07 22:21:14 crc kubenswrapper[4871]: E1007 22:21:14.809019 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0609d7b1-7443-4bcd-935e-ca2a61a359b7" containerName="controller-manager" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.809037 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0609d7b1-7443-4bcd-935e-ca2a61a359b7" containerName="controller-manager" Oct 07 22:21:14 crc kubenswrapper[4871]: E1007 22:21:14.809070 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ce8147-0273-421d-9263-3d4eb06bfcfe" containerName="route-controller-manager" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.809088 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ce8147-0273-421d-9263-3d4eb06bfcfe" containerName="route-controller-manager" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.809340 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0609d7b1-7443-4bcd-935e-ca2a61a359b7" containerName="controller-manager" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.809374 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="03ce8147-0273-421d-9263-3d4eb06bfcfe" containerName="route-controller-manager" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.809403 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3689423d-c1f1-46c8-9ea5-165147a8a286" containerName="registry" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.810476 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.815257 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.815758 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.816054 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.816921 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.817084 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.817230 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.831862 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-544dd56fb8-dj9qr"] Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.846838 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.857119 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf"] Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.861817 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.862716 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.864010 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.864133 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.871752 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.873518 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13bccc88-4e1d-4045-9c9b-8afac5f72027-serving-cert\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.873803 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13bccc88-4e1d-4045-9c9b-8afac5f72027-client-ca\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.873854 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrbbn\" (UniqueName: \"kubernetes.io/projected/13bccc88-4e1d-4045-9c9b-8afac5f72027-kube-api-access-zrbbn\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.873944 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13bccc88-4e1d-4045-9c9b-8afac5f72027-config\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.874985 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.883852 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-544dd56fb8-dj9qr"] Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.885954 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.975411 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c780db-2066-4e44-accd-3e1f79f6f055-serving-cert\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.975475 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjd7s\" (UniqueName: \"kubernetes.io/projected/05c780db-2066-4e44-accd-3e1f79f6f055-kube-api-access-gjd7s\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.975510 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05c780db-2066-4e44-accd-3e1f79f6f055-proxy-ca-bundles\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.975539 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13bccc88-4e1d-4045-9c9b-8afac5f72027-client-ca\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.975574 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrbbn\" (UniqueName: \"kubernetes.io/projected/13bccc88-4e1d-4045-9c9b-8afac5f72027-kube-api-access-zrbbn\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.975600 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05c780db-2066-4e44-accd-3e1f79f6f055-client-ca\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.975699 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13bccc88-4e1d-4045-9c9b-8afac5f72027-config\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.975762 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13bccc88-4e1d-4045-9c9b-8afac5f72027-serving-cert\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.975809 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c780db-2066-4e44-accd-3e1f79f6f055-config\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.976938 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13bccc88-4e1d-4045-9c9b-8afac5f72027-client-ca\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.978347 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13bccc88-4e1d-4045-9c9b-8afac5f72027-config\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.983861 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13bccc88-4e1d-4045-9c9b-8afac5f72027-serving-cert\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.989720 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03ce8147-0273-421d-9263-3d4eb06bfcfe" path="/var/lib/kubelet/pods/03ce8147-0273-421d-9263-3d4eb06bfcfe/volumes" Oct 07 22:21:14 crc kubenswrapper[4871]: I1007 22:21:14.990954 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0609d7b1-7443-4bcd-935e-ca2a61a359b7" path="/var/lib/kubelet/pods/0609d7b1-7443-4bcd-935e-ca2a61a359b7/volumes" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.003086 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrbbn\" (UniqueName: \"kubernetes.io/projected/13bccc88-4e1d-4045-9c9b-8afac5f72027-kube-api-access-zrbbn\") pod \"route-controller-manager-8b4b46c96-srrlf\" (UID: \"13bccc88-4e1d-4045-9c9b-8afac5f72027\") " pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.077721 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjd7s\" (UniqueName: \"kubernetes.io/projected/05c780db-2066-4e44-accd-3e1f79f6f055-kube-api-access-gjd7s\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.077785 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05c780db-2066-4e44-accd-3e1f79f6f055-proxy-ca-bundles\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.077831 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05c780db-2066-4e44-accd-3e1f79f6f055-client-ca\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.077890 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c780db-2066-4e44-accd-3e1f79f6f055-config\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.077925 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c780db-2066-4e44-accd-3e1f79f6f055-serving-cert\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.079170 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/05c780db-2066-4e44-accd-3e1f79f6f055-proxy-ca-bundles\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.079404 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c780db-2066-4e44-accd-3e1f79f6f055-config\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.080132 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/05c780db-2066-4e44-accd-3e1f79f6f055-client-ca\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.083102 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c780db-2066-4e44-accd-3e1f79f6f055-serving-cert\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.101782 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjd7s\" (UniqueName: \"kubernetes.io/projected/05c780db-2066-4e44-accd-3e1f79f6f055-kube-api-access-gjd7s\") pod \"controller-manager-544dd56fb8-dj9qr\" (UID: \"05c780db-2066-4e44-accd-3e1f79f6f055\") " pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.154819 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.210160 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.443734 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf"] Oct 07 22:21:15 crc kubenswrapper[4871]: I1007 22:21:15.520417 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-544dd56fb8-dj9qr"] Oct 07 22:21:15 crc kubenswrapper[4871]: W1007 22:21:15.531379 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05c780db_2066_4e44_accd_3e1f79f6f055.slice/crio-d2889a497f5e4d018d8ac45f74c7c43f9a0ffe2c131684d06ad824e0dbcbf5f4 WatchSource:0}: Error finding container d2889a497f5e4d018d8ac45f74c7c43f9a0ffe2c131684d06ad824e0dbcbf5f4: Status 404 returned error can't find the container with id d2889a497f5e4d018d8ac45f74c7c43f9a0ffe2c131684d06ad824e0dbcbf5f4 Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.131353 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" event={"ID":"13bccc88-4e1d-4045-9c9b-8afac5f72027","Type":"ContainerStarted","Data":"c0bf11c0edef6a8231b1985ab809fed346cd1eeb5b2b32eab5adf3cc86266bb6"} Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.131717 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.131731 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" event={"ID":"13bccc88-4e1d-4045-9c9b-8afac5f72027","Type":"ContainerStarted","Data":"9244cd56a74a0659854de34d0bd34bbc9940b8153021901132f3163788f7749b"} Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.133101 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" event={"ID":"05c780db-2066-4e44-accd-3e1f79f6f055","Type":"ContainerStarted","Data":"3f82089ba91d66435f56040fb6712415593f3da0a26cbaa7abb8b0f8a00b0872"} Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.133158 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" event={"ID":"05c780db-2066-4e44-accd-3e1f79f6f055","Type":"ContainerStarted","Data":"d2889a497f5e4d018d8ac45f74c7c43f9a0ffe2c131684d06ad824e0dbcbf5f4"} Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.133180 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.136398 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.138119 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.156757 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8b4b46c96-srrlf" podStartSLOduration=3.156731141 podStartE2EDuration="3.156731141s" podCreationTimestamp="2025-10-07 22:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:21:16.155580781 +0000 UTC m=+749.958278854" watchObservedRunningTime="2025-10-07 22:21:16.156731141 +0000 UTC m=+749.959429214" Oct 07 22:21:16 crc kubenswrapper[4871]: I1007 22:21:16.188050 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-544dd56fb8-dj9qr" podStartSLOduration=3.18802949 podStartE2EDuration="3.18802949s" podCreationTimestamp="2025-10-07 22:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:21:16.184913568 +0000 UTC m=+749.987611641" watchObservedRunningTime="2025-10-07 22:21:16.18802949 +0000 UTC m=+749.990727563" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.426347 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-d44j5"] Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.428128 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="nbdb" containerID="cri-o://01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" gracePeriod=30 Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.428161 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737" gracePeriod=30 Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.428199 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="northd" containerID="cri-o://a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564" gracePeriod=30 Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.428238 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kube-rbac-proxy-node" containerID="cri-o://2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8" gracePeriod=30 Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.428286 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovn-acl-logging" containerID="cri-o://627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e" gracePeriod=30 Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.428368 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="sbdb" containerID="cri-o://dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" gracePeriod=30 Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.428441 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovn-controller" containerID="cri-o://92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70" gracePeriod=30 Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.470812 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.473065 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.473276 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.475540 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.475752 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.475899 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="nbdb" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.484359 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.484417 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="sbdb" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.484564 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" containerID="cri-o://c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf" gracePeriod=30 Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.785277 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/3.log" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.791533 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovn-acl-logging/0.log" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.792179 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovn-controller/0.log" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.792748 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847065 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8h4fg"] Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847359 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847384 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847394 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847401 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847410 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="northd" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847418 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="northd" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847434 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="sbdb" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847442 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="sbdb" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847455 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kubecfg-setup" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847464 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kubecfg-setup" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847476 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovn-acl-logging" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847485 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovn-acl-logging" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847497 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovn-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847504 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovn-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847518 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kube-rbac-proxy-node" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847525 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kube-rbac-proxy-node" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847534 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847542 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847554 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847561 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847571 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847578 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847590 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="nbdb" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847597 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="nbdb" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847692 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="nbdb" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847707 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="sbdb" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847713 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847719 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847730 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovn-acl-logging" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847739 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847748 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847757 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="kube-rbac-proxy-node" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847766 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovn-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847774 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="northd" Oct 07 22:21:19 crc kubenswrapper[4871]: E1007 22:21:19.847885 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847895 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.847983 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.848172 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerName="ovnkube-controller" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.849633 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.852758 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-netd\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.852841 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-var-lib-openvswitch\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.852897 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.852904 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-config\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853013 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-etc-openvswitch\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853063 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7hp6\" (UniqueName: \"kubernetes.io/projected/0059e7e8-7d63-4b03-81a0-b8521803f34a-kube-api-access-q7hp6\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853107 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-env-overrides\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853143 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-systemd-units\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853177 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-log-socket\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853227 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853257 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-systemd\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853285 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-bin\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853343 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-kubelet\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853393 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-script-lib\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853442 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-ovn-kubernetes\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853494 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-node-log\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853548 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovn-node-metrics-cert\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853572 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-slash\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853584 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853614 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-ovn\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853634 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853642 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-netns\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853669 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853674 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-openvswitch\") pod \"0059e7e8-7d63-4b03-81a0-b8521803f34a\" (UID: \"0059e7e8-7d63-4b03-81a0-b8521803f34a\") " Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853707 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.853750 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854240 4871 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854274 4871 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854286 4871 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854298 4871 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854311 4871 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854309 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854326 4871 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854358 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854363 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854387 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-log-socket" (OuterVolumeSpecName: "log-socket") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854411 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-node-log" (OuterVolumeSpecName: "node-log") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854420 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.854426 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.855608 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.855660 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-slash" (OuterVolumeSpecName: "host-slash") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.855691 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.856024 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.861941 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.862432 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0059e7e8-7d63-4b03-81a0-b8521803f34a-kube-api-access-q7hp6" (OuterVolumeSpecName: "kube-api-access-q7hp6") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "kube-api-access-q7hp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.870905 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "0059e7e8-7d63-4b03-81a0-b8521803f34a" (UID: "0059e7e8-7d63-4b03-81a0-b8521803f34a"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.955569 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-run-ovn\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.955624 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8f62d023-6b98-4932-ae14-c50e7f538a4e-ovnkube-config\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.955649 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-run-ovn-kubernetes\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.955753 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-run-netns\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.955941 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-cni-bin\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.955966 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-systemd-units\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.955989 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-kubelet\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956055 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t6d9\" (UniqueName: \"kubernetes.io/projected/8f62d023-6b98-4932-ae14-c50e7f538a4e-kube-api-access-2t6d9\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956093 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8f62d023-6b98-4932-ae14-c50e7f538a4e-ovn-node-metrics-cert\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956123 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-slash\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956301 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-run-openvswitch\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956422 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-etc-openvswitch\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956482 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-node-log\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956517 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-run-systemd\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956547 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-log-socket\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956841 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956905 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-var-lib-openvswitch\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956935 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8f62d023-6b98-4932-ae14-c50e7f538a4e-ovnkube-script-lib\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956963 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8f62d023-6b98-4932-ae14-c50e7f538a4e-env-overrides\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.956997 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-cni-netd\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957103 4871 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957120 4871 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957136 4871 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957148 4871 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957156 4871 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-node-log\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957166 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0059e7e8-7d63-4b03-81a0-b8521803f34a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957175 4871 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-slash\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957184 4871 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957193 4871 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957203 4871 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957213 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7hp6\" (UniqueName: \"kubernetes.io/projected/0059e7e8-7d63-4b03-81a0-b8521803f34a-kube-api-access-q7hp6\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957223 4871 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0059e7e8-7d63-4b03-81a0-b8521803f34a-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957231 4871 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-log-socket\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:19 crc kubenswrapper[4871]: I1007 22:21:19.957240 4871 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0059e7e8-7d63-4b03-81a0-b8521803f34a-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.059874 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-var-lib-openvswitch\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.059921 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.059947 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8f62d023-6b98-4932-ae14-c50e7f538a4e-ovnkube-script-lib\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.059968 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8f62d023-6b98-4932-ae14-c50e7f538a4e-env-overrides\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.059990 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-cni-netd\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060006 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-run-ovn\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060023 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8f62d023-6b98-4932-ae14-c50e7f538a4e-ovnkube-config\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060032 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-var-lib-openvswitch\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060076 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-run-ovn-kubernetes\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060047 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-run-ovn-kubernetes\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060141 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-run-netns\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060192 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-cni-bin\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060045 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060219 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-systemd-units\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060254 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-run-netns\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060281 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-cni-netd\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060287 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-kubelet\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060338 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t6d9\" (UniqueName: \"kubernetes.io/projected/8f62d023-6b98-4932-ae14-c50e7f538a4e-kube-api-access-2t6d9\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060379 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8f62d023-6b98-4932-ae14-c50e7f538a4e-ovn-node-metrics-cert\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060422 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-slash\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060479 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-run-openvswitch\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060525 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-etc-openvswitch\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060548 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-node-log\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060572 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-run-systemd\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060597 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-log-socket\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060625 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-run-ovn\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060838 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8f62d023-6b98-4932-ae14-c50e7f538a4e-env-overrides\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060855 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-slash\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060863 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-log-socket\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060869 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-systemd-units\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060881 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-etc-openvswitch\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060952 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-cni-bin\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060908 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-node-log\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060919 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-host-kubelet\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060930 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-run-systemd\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.060903 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8f62d023-6b98-4932-ae14-c50e7f538a4e-run-openvswitch\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.061511 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8f62d023-6b98-4932-ae14-c50e7f538a4e-ovnkube-config\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.063233 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8f62d023-6b98-4932-ae14-c50e7f538a4e-ovnkube-script-lib\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.063905 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8f62d023-6b98-4932-ae14-c50e7f538a4e-ovn-node-metrics-cert\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.086886 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t6d9\" (UniqueName: \"kubernetes.io/projected/8f62d023-6b98-4932-ae14-c50e7f538a4e-kube-api-access-2t6d9\") pod \"ovnkube-node-8h4fg\" (UID: \"8f62d023-6b98-4932-ae14-c50e7f538a4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.168131 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.173171 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nv4jc_53708429-59b4-4319-bdb7-8a922a551e59/kube-multus/2.log" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.173830 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nv4jc_53708429-59b4-4319-bdb7-8a922a551e59/kube-multus/1.log" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.173890 4871 generic.go:334] "Generic (PLEG): container finished" podID="53708429-59b4-4319-bdb7-8a922a551e59" containerID="603cedf2bf5b244f3d25d0cc6a1e215afab476ce358d9dfdef028d22f6c3f899" exitCode=2 Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.173972 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nv4jc" event={"ID":"53708429-59b4-4319-bdb7-8a922a551e59","Type":"ContainerDied","Data":"603cedf2bf5b244f3d25d0cc6a1e215afab476ce358d9dfdef028d22f6c3f899"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.174033 4871 scope.go:117] "RemoveContainer" containerID="757c570312a99f9653e7c2d511e8d97a7cff614d542e78c392b717f079282d3d" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.175694 4871 scope.go:117] "RemoveContainer" containerID="603cedf2bf5b244f3d25d0cc6a1e215afab476ce358d9dfdef028d22f6c3f899" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.180088 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovnkube-controller/3.log" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.188155 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovn-acl-logging/0.log" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.188682 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d44j5_0059e7e8-7d63-4b03-81a0-b8521803f34a/ovn-controller/0.log" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189164 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf" exitCode=0 Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189187 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" exitCode=0 Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189195 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" exitCode=0 Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189204 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564" exitCode=0 Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189218 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189212 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737" exitCode=0 Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189238 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8" exitCode=0 Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189245 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e" exitCode=143 Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189252 4871 generic.go:334] "Generic (PLEG): container finished" podID="0059e7e8-7d63-4b03-81a0-b8521803f34a" containerID="92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70" exitCode=143 Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189250 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189337 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189353 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189371 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189382 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189399 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189413 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189431 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189439 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189445 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189451 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189458 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189464 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189471 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189477 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189485 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189493 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189504 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189512 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189517 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189525 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189531 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189537 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189545 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189551 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189556 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189561 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189570 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189580 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189588 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189595 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189601 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189607 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189613 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189619 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189627 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189633 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189639 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189648 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d44j5" event={"ID":"0059e7e8-7d63-4b03-81a0-b8521803f34a","Type":"ContainerDied","Data":"f24c5e6ebbe3f07e4af9c5d1a9a86bdff3fd1059ba514824a6b16f188d5c668b"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189660 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189668 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189674 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189680 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189687 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189694 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189701 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189708 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189716 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.189724 4871 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd"} Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.221178 4871 scope.go:117] "RemoveContainer" containerID="c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.239681 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-d44j5"] Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.244358 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-d44j5"] Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.246507 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.279744 4871 scope.go:117] "RemoveContainer" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.305107 4871 scope.go:117] "RemoveContainer" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.327061 4871 scope.go:117] "RemoveContainer" containerID="a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.345144 4871 scope.go:117] "RemoveContainer" containerID="2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.362165 4871 scope.go:117] "RemoveContainer" containerID="2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.390881 4871 scope.go:117] "RemoveContainer" containerID="627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.444485 4871 scope.go:117] "RemoveContainer" containerID="92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.474163 4871 scope.go:117] "RemoveContainer" containerID="376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.490861 4871 scope.go:117] "RemoveContainer" containerID="c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.491469 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf\": container with ID starting with c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf not found: ID does not exist" containerID="c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.491527 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} err="failed to get container status \"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf\": rpc error: code = NotFound desc = could not find container \"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf\": container with ID starting with c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.491564 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.493515 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\": container with ID starting with a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f not found: ID does not exist" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.493556 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} err="failed to get container status \"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\": rpc error: code = NotFound desc = could not find container \"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\": container with ID starting with a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.493606 4871 scope.go:117] "RemoveContainer" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.494162 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\": container with ID starting with dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab not found: ID does not exist" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.494211 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} err="failed to get container status \"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\": rpc error: code = NotFound desc = could not find container \"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\": container with ID starting with dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.494249 4871 scope.go:117] "RemoveContainer" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.494664 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\": container with ID starting with 01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41 not found: ID does not exist" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.494685 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} err="failed to get container status \"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\": rpc error: code = NotFound desc = could not find container \"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\": container with ID starting with 01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.494699 4871 scope.go:117] "RemoveContainer" containerID="a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.495056 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\": container with ID starting with a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564 not found: ID does not exist" containerID="a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.495107 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} err="failed to get container status \"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\": rpc error: code = NotFound desc = could not find container \"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\": container with ID starting with a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.495147 4871 scope.go:117] "RemoveContainer" containerID="2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.495738 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\": container with ID starting with 2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737 not found: ID does not exist" containerID="2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.495863 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} err="failed to get container status \"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\": rpc error: code = NotFound desc = could not find container \"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\": container with ID starting with 2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.495890 4871 scope.go:117] "RemoveContainer" containerID="2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.496592 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\": container with ID starting with 2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8 not found: ID does not exist" containerID="2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.496623 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} err="failed to get container status \"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\": rpc error: code = NotFound desc = could not find container \"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\": container with ID starting with 2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.496645 4871 scope.go:117] "RemoveContainer" containerID="627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.497027 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\": container with ID starting with 627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e not found: ID does not exist" containerID="627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.497054 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} err="failed to get container status \"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\": rpc error: code = NotFound desc = could not find container \"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\": container with ID starting with 627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.497073 4871 scope.go:117] "RemoveContainer" containerID="92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.497366 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\": container with ID starting with 92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70 not found: ID does not exist" containerID="92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.497404 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} err="failed to get container status \"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\": rpc error: code = NotFound desc = could not find container \"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\": container with ID starting with 92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.497425 4871 scope.go:117] "RemoveContainer" containerID="376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd" Oct 07 22:21:20 crc kubenswrapper[4871]: E1007 22:21:20.497718 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\": container with ID starting with 376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd not found: ID does not exist" containerID="376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.497745 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd"} err="failed to get container status \"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\": rpc error: code = NotFound desc = could not find container \"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\": container with ID starting with 376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.497761 4871 scope.go:117] "RemoveContainer" containerID="c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.498132 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} err="failed to get container status \"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf\": rpc error: code = NotFound desc = could not find container \"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf\": container with ID starting with c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.498161 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.498458 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} err="failed to get container status \"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\": rpc error: code = NotFound desc = could not find container \"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\": container with ID starting with a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.498479 4871 scope.go:117] "RemoveContainer" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.498745 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} err="failed to get container status \"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\": rpc error: code = NotFound desc = could not find container \"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\": container with ID starting with dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.498767 4871 scope.go:117] "RemoveContainer" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.499051 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} err="failed to get container status \"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\": rpc error: code = NotFound desc = could not find container \"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\": container with ID starting with 01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.499072 4871 scope.go:117] "RemoveContainer" containerID="a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.499443 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} err="failed to get container status \"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\": rpc error: code = NotFound desc = could not find container \"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\": container with ID starting with a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.499464 4871 scope.go:117] "RemoveContainer" containerID="2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.499925 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} err="failed to get container status \"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\": rpc error: code = NotFound desc = could not find container \"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\": container with ID starting with 2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.499948 4871 scope.go:117] "RemoveContainer" containerID="2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.500311 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} err="failed to get container status \"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\": rpc error: code = NotFound desc = could not find container \"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\": container with ID starting with 2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.500329 4871 scope.go:117] "RemoveContainer" containerID="627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.500630 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} err="failed to get container status \"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\": rpc error: code = NotFound desc = could not find container \"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\": container with ID starting with 627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.500656 4871 scope.go:117] "RemoveContainer" containerID="92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.500989 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} err="failed to get container status \"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\": rpc error: code = NotFound desc = could not find container \"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\": container with ID starting with 92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.501010 4871 scope.go:117] "RemoveContainer" containerID="376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.501391 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd"} err="failed to get container status \"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\": rpc error: code = NotFound desc = could not find container \"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\": container with ID starting with 376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.501433 4871 scope.go:117] "RemoveContainer" containerID="c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.501819 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} err="failed to get container status \"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf\": rpc error: code = NotFound desc = could not find container \"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf\": container with ID starting with c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.501852 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.502295 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} err="failed to get container status \"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\": rpc error: code = NotFound desc = could not find container \"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\": container with ID starting with a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.502314 4871 scope.go:117] "RemoveContainer" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.502624 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} err="failed to get container status \"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\": rpc error: code = NotFound desc = could not find container \"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\": container with ID starting with dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.502652 4871 scope.go:117] "RemoveContainer" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.503154 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} err="failed to get container status \"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\": rpc error: code = NotFound desc = could not find container \"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\": container with ID starting with 01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.503181 4871 scope.go:117] "RemoveContainer" containerID="a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.503527 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} err="failed to get container status \"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\": rpc error: code = NotFound desc = could not find container \"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\": container with ID starting with a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.503560 4871 scope.go:117] "RemoveContainer" containerID="2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.504102 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} err="failed to get container status \"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\": rpc error: code = NotFound desc = could not find container \"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\": container with ID starting with 2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.504138 4871 scope.go:117] "RemoveContainer" containerID="2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.504409 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} err="failed to get container status \"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\": rpc error: code = NotFound desc = could not find container \"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\": container with ID starting with 2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.504451 4871 scope.go:117] "RemoveContainer" containerID="627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.504910 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} err="failed to get container status \"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\": rpc error: code = NotFound desc = could not find container \"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\": container with ID starting with 627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.504932 4871 scope.go:117] "RemoveContainer" containerID="92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.505197 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} err="failed to get container status \"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\": rpc error: code = NotFound desc = could not find container \"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\": container with ID starting with 92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.505220 4871 scope.go:117] "RemoveContainer" containerID="376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.505696 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd"} err="failed to get container status \"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\": rpc error: code = NotFound desc = could not find container \"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\": container with ID starting with 376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.505722 4871 scope.go:117] "RemoveContainer" containerID="c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.506042 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf"} err="failed to get container status \"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf\": rpc error: code = NotFound desc = could not find container \"c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf\": container with ID starting with c1be4a65a016d40e97239b0d9ee68da92467a36303ba90f25f865c2c61af78bf not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.506069 4871 scope.go:117] "RemoveContainer" containerID="a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.506339 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f"} err="failed to get container status \"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\": rpc error: code = NotFound desc = could not find container \"a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f\": container with ID starting with a70411e61c809048417523fcd4505f30e583890e476ecfc8a86174cd391f965f not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.506379 4871 scope.go:117] "RemoveContainer" containerID="dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.506876 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab"} err="failed to get container status \"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\": rpc error: code = NotFound desc = could not find container \"dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab\": container with ID starting with dcd5ed0a79ad8953ce90dda2e9243715ab9d34ff5257d7aba4368ccaf35073ab not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.506954 4871 scope.go:117] "RemoveContainer" containerID="01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.507532 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41"} err="failed to get container status \"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\": rpc error: code = NotFound desc = could not find container \"01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41\": container with ID starting with 01ac8566aa397e3b649213e29ebce12a7753fd1b49fce7eeb8514d77b00c9d41 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.507562 4871 scope.go:117] "RemoveContainer" containerID="a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.508024 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564"} err="failed to get container status \"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\": rpc error: code = NotFound desc = could not find container \"a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564\": container with ID starting with a42766426957057bf009c0b0ed8f7f2e75386d624ccebf822bfbcc93b6070564 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.508056 4871 scope.go:117] "RemoveContainer" containerID="2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.508371 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737"} err="failed to get container status \"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\": rpc error: code = NotFound desc = could not find container \"2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737\": container with ID starting with 2dc23aa963502dc71a704d2b0bb65261902c8c51e169977f3869a6a4a3a68737 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.508410 4871 scope.go:117] "RemoveContainer" containerID="2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.508809 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8"} err="failed to get container status \"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\": rpc error: code = NotFound desc = could not find container \"2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8\": container with ID starting with 2f928a7357e4ee083a7164f7be9f7f7b551e60c973de502c57881b7b5d92d7f8 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.508831 4871 scope.go:117] "RemoveContainer" containerID="627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.509139 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e"} err="failed to get container status \"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\": rpc error: code = NotFound desc = could not find container \"627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e\": container with ID starting with 627621a9e0f756bce5baa95dce0af0060d930f0c4a3148679dac04ab7507480e not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.509165 4871 scope.go:117] "RemoveContainer" containerID="92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.509479 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70"} err="failed to get container status \"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\": rpc error: code = NotFound desc = could not find container \"92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70\": container with ID starting with 92bb86d48e273e9eebb6d323a734af7b46f15424c12bf0ebd81941014b57fe70 not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.509504 4871 scope.go:117] "RemoveContainer" containerID="376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.509769 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd"} err="failed to get container status \"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\": rpc error: code = NotFound desc = could not find container \"376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd\": container with ID starting with 376f721d564748a37f8106a78b66b98102efbeb3cf516ab2b123ccaf264f3fdd not found: ID does not exist" Oct 07 22:21:20 crc kubenswrapper[4871]: I1007 22:21:20.990141 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0059e7e8-7d63-4b03-81a0-b8521803f34a" path="/var/lib/kubelet/pods/0059e7e8-7d63-4b03-81a0-b8521803f34a/volumes" Oct 07 22:21:21 crc kubenswrapper[4871]: I1007 22:21:21.197389 4871 generic.go:334] "Generic (PLEG): container finished" podID="8f62d023-6b98-4932-ae14-c50e7f538a4e" containerID="01cac5563e5d38cfc0aa2c21d7a19dc1b7117ac1faabfa1f182df76714befb86" exitCode=0 Oct 07 22:21:21 crc kubenswrapper[4871]: I1007 22:21:21.197452 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerDied","Data":"01cac5563e5d38cfc0aa2c21d7a19dc1b7117ac1faabfa1f182df76714befb86"} Oct 07 22:21:21 crc kubenswrapper[4871]: I1007 22:21:21.197514 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerStarted","Data":"fef5c5808c769da35ac86003706821b53cb49173436352003e1975225265d667"} Oct 07 22:21:21 crc kubenswrapper[4871]: I1007 22:21:21.204943 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nv4jc_53708429-59b4-4319-bdb7-8a922a551e59/kube-multus/2.log" Oct 07 22:21:21 crc kubenswrapper[4871]: I1007 22:21:21.205012 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nv4jc" event={"ID":"53708429-59b4-4319-bdb7-8a922a551e59","Type":"ContainerStarted","Data":"c0cfd93ca484a67a0bc7e3fc081918df31ce1758adc6d896b92e333a5e199a5c"} Oct 07 22:21:21 crc kubenswrapper[4871]: I1007 22:21:21.851445 4871 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 22:21:22 crc kubenswrapper[4871]: I1007 22:21:22.216435 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerStarted","Data":"e6a3c4517bd5d4adc76d0358863dfe856e4565436909582f45aaaece91a167b7"} Oct 07 22:21:22 crc kubenswrapper[4871]: I1007 22:21:22.216816 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerStarted","Data":"99fbad28a9b63ca891493ee488547f9190eb6bce722b2007ecebb21d021801d4"} Oct 07 22:21:22 crc kubenswrapper[4871]: I1007 22:21:22.216829 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerStarted","Data":"f2dbd55642ad9c069c400d0a080b49f5db87f987dc3cedcf80f7364d98f3c80d"} Oct 07 22:21:22 crc kubenswrapper[4871]: I1007 22:21:22.216838 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerStarted","Data":"a30ff9500067da9395bb44895030ca2887b5d7aa2d405a91b4c1f296b2ef5df8"} Oct 07 22:21:22 crc kubenswrapper[4871]: I1007 22:21:22.216846 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerStarted","Data":"ea94fb7aeb1409db4b18f078af108a6fec152003b3cc8347b28a630010b4f136"} Oct 07 22:21:22 crc kubenswrapper[4871]: I1007 22:21:22.216856 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerStarted","Data":"5868df348b2fc75a3b12596ed6543f4c591e5ea71dea05b9a7d12b3c6e897713"} Oct 07 22:21:25 crc kubenswrapper[4871]: I1007 22:21:25.240504 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerStarted","Data":"2cad2bcc9e872ef5020f8638bb1b20acf55a64d59d97ecef40e7186f40856b48"} Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.031090 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-hntcc"] Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.033034 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.036782 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.036887 4871 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-595tx" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.036886 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.038027 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.062822 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbbvm\" (UniqueName: \"kubernetes.io/projected/d7cd42c4-268b-4588-a427-0a80502fe808-kube-api-access-hbbvm\") pod \"crc-storage-crc-hntcc\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.062936 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d7cd42c4-268b-4588-a427-0a80502fe808-node-mnt\") pod \"crc-storage-crc-hntcc\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.062982 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d7cd42c4-268b-4588-a427-0a80502fe808-crc-storage\") pod \"crc-storage-crc-hntcc\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.163741 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbbvm\" (UniqueName: \"kubernetes.io/projected/d7cd42c4-268b-4588-a427-0a80502fe808-kube-api-access-hbbvm\") pod \"crc-storage-crc-hntcc\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.164481 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d7cd42c4-268b-4588-a427-0a80502fe808-node-mnt\") pod \"crc-storage-crc-hntcc\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.164512 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d7cd42c4-268b-4588-a427-0a80502fe808-crc-storage\") pod \"crc-storage-crc-hntcc\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.164954 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d7cd42c4-268b-4588-a427-0a80502fe808-node-mnt\") pod \"crc-storage-crc-hntcc\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.165336 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d7cd42c4-268b-4588-a427-0a80502fe808-crc-storage\") pod \"crc-storage-crc-hntcc\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.189322 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbbvm\" (UniqueName: \"kubernetes.io/projected/d7cd42c4-268b-4588-a427-0a80502fe808-kube-api-access-hbbvm\") pod \"crc-storage-crc-hntcc\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.278348 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" event={"ID":"8f62d023-6b98-4932-ae14-c50e7f538a4e","Type":"ContainerStarted","Data":"12ca7acfacb80283a8290196a223e8f063a003adef84c45ad06379d06e4827f8"} Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.279266 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.279326 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.279343 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.317127 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.330973 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.366361 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.374995 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" podStartSLOduration=8.37496454 podStartE2EDuration="8.37496454s" podCreationTimestamp="2025-10-07 22:21:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:21:27.329134666 +0000 UTC m=+761.131832759" watchObservedRunningTime="2025-10-07 22:21:27.37496454 +0000 UTC m=+761.177662613" Oct 07 22:21:27 crc kubenswrapper[4871]: E1007 22:21:27.402389 4871 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-hntcc_crc-storage_d7cd42c4-268b-4588-a427-0a80502fe808_0(18b468dc3da7e58c72483cb3d680da7622a3de0953f7558d4ca5f8fcdd00e35c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 22:21:27 crc kubenswrapper[4871]: E1007 22:21:27.402563 4871 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-hntcc_crc-storage_d7cd42c4-268b-4588-a427-0a80502fe808_0(18b468dc3da7e58c72483cb3d680da7622a3de0953f7558d4ca5f8fcdd00e35c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: E1007 22:21:27.402605 4871 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-hntcc_crc-storage_d7cd42c4-268b-4588-a427-0a80502fe808_0(18b468dc3da7e58c72483cb3d680da7622a3de0953f7558d4ca5f8fcdd00e35c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:27 crc kubenswrapper[4871]: E1007 22:21:27.403642 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-hntcc_crc-storage(d7cd42c4-268b-4588-a427-0a80502fe808)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-hntcc_crc-storage(d7cd42c4-268b-4588-a427-0a80502fe808)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-hntcc_crc-storage_d7cd42c4-268b-4588-a427-0a80502fe808_0(18b468dc3da7e58c72483cb3d680da7622a3de0953f7558d4ca5f8fcdd00e35c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-hntcc" podUID="d7cd42c4-268b-4588-a427-0a80502fe808" Oct 07 22:21:27 crc kubenswrapper[4871]: I1007 22:21:27.884246 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-hntcc"] Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.283024 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.284050 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:28 crc kubenswrapper[4871]: E1007 22:21:28.307230 4871 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-hntcc_crc-storage_d7cd42c4-268b-4588-a427-0a80502fe808_0(f70e73cda4dc91d0c4399fac95a453a01c6cdd3a18d133e287e9f39ede2d7dcc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 22:21:28 crc kubenswrapper[4871]: E1007 22:21:28.307308 4871 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-hntcc_crc-storage_d7cd42c4-268b-4588-a427-0a80502fe808_0(f70e73cda4dc91d0c4399fac95a453a01c6cdd3a18d133e287e9f39ede2d7dcc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:28 crc kubenswrapper[4871]: E1007 22:21:28.307338 4871 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-hntcc_crc-storage_d7cd42c4-268b-4588-a427-0a80502fe808_0(f70e73cda4dc91d0c4399fac95a453a01c6cdd3a18d133e287e9f39ede2d7dcc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:28 crc kubenswrapper[4871]: E1007 22:21:28.307406 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-hntcc_crc-storage(d7cd42c4-268b-4588-a427-0a80502fe808)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-hntcc_crc-storage(d7cd42c4-268b-4588-a427-0a80502fe808)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-hntcc_crc-storage_d7cd42c4-268b-4588-a427-0a80502fe808_0(f70e73cda4dc91d0c4399fac95a453a01c6cdd3a18d133e287e9f39ede2d7dcc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-hntcc" podUID="d7cd42c4-268b-4588-a427-0a80502fe808" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.716713 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q47sb"] Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.718572 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.724305 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q47sb"] Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.797409 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-catalog-content\") pod \"redhat-marketplace-q47sb\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.797485 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-utilities\") pod \"redhat-marketplace-q47sb\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.797601 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kkkw\" (UniqueName: \"kubernetes.io/projected/6003e812-ee55-4cf6-b265-fe3f242d7fab-kube-api-access-9kkkw\") pod \"redhat-marketplace-q47sb\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.899580 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-catalog-content\") pod \"redhat-marketplace-q47sb\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.899665 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-utilities\") pod \"redhat-marketplace-q47sb\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.899845 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kkkw\" (UniqueName: \"kubernetes.io/projected/6003e812-ee55-4cf6-b265-fe3f242d7fab-kube-api-access-9kkkw\") pod \"redhat-marketplace-q47sb\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.900434 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-catalog-content\") pod \"redhat-marketplace-q47sb\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.900494 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-utilities\") pod \"redhat-marketplace-q47sb\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:28 crc kubenswrapper[4871]: I1007 22:21:28.921733 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kkkw\" (UniqueName: \"kubernetes.io/projected/6003e812-ee55-4cf6-b265-fe3f242d7fab-kube-api-access-9kkkw\") pod \"redhat-marketplace-q47sb\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:29 crc kubenswrapper[4871]: I1007 22:21:29.047982 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:29 crc kubenswrapper[4871]: E1007 22:21:29.069266 4871 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-q47sb_openshift-marketplace_6003e812-ee55-4cf6-b265-fe3f242d7fab_0(ccf0c64906c099878abfa56e948a2d30ad3263305c531b03a08a1733a9272cdd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 22:21:29 crc kubenswrapper[4871]: E1007 22:21:29.069377 4871 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-q47sb_openshift-marketplace_6003e812-ee55-4cf6-b265-fe3f242d7fab_0(ccf0c64906c099878abfa56e948a2d30ad3263305c531b03a08a1733a9272cdd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:29 crc kubenswrapper[4871]: E1007 22:21:29.069422 4871 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-q47sb_openshift-marketplace_6003e812-ee55-4cf6-b265-fe3f242d7fab_0(ccf0c64906c099878abfa56e948a2d30ad3263305c531b03a08a1733a9272cdd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:29 crc kubenswrapper[4871]: E1007 22:21:29.069524 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-marketplace-q47sb_openshift-marketplace(6003e812-ee55-4cf6-b265-fe3f242d7fab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-marketplace-q47sb_openshift-marketplace(6003e812-ee55-4cf6-b265-fe3f242d7fab)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-q47sb_openshift-marketplace_6003e812-ee55-4cf6-b265-fe3f242d7fab_0(ccf0c64906c099878abfa56e948a2d30ad3263305c531b03a08a1733a9272cdd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-marketplace-q47sb" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" Oct 07 22:21:29 crc kubenswrapper[4871]: I1007 22:21:29.287917 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:29 crc kubenswrapper[4871]: I1007 22:21:29.288829 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:29 crc kubenswrapper[4871]: E1007 22:21:29.311195 4871 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-q47sb_openshift-marketplace_6003e812-ee55-4cf6-b265-fe3f242d7fab_0(4139941265c5376d77713744d0f0848fcfcff7d8faf533809faf669eeda18a96): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 22:21:29 crc kubenswrapper[4871]: E1007 22:21:29.311297 4871 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-q47sb_openshift-marketplace_6003e812-ee55-4cf6-b265-fe3f242d7fab_0(4139941265c5376d77713744d0f0848fcfcff7d8faf533809faf669eeda18a96): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:29 crc kubenswrapper[4871]: E1007 22:21:29.311326 4871 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-q47sb_openshift-marketplace_6003e812-ee55-4cf6-b265-fe3f242d7fab_0(4139941265c5376d77713744d0f0848fcfcff7d8faf533809faf669eeda18a96): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:29 crc kubenswrapper[4871]: E1007 22:21:29.311382 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-marketplace-q47sb_openshift-marketplace(6003e812-ee55-4cf6-b265-fe3f242d7fab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-marketplace-q47sb_openshift-marketplace(6003e812-ee55-4cf6-b265-fe3f242d7fab)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-q47sb_openshift-marketplace_6003e812-ee55-4cf6-b265-fe3f242d7fab_0(4139941265c5376d77713744d0f0848fcfcff7d8faf533809faf669eeda18a96): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-marketplace-q47sb" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" Oct 07 22:21:35 crc kubenswrapper[4871]: I1007 22:21:35.512399 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:21:35 crc kubenswrapper[4871]: I1007 22:21:35.513416 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:21:35 crc kubenswrapper[4871]: I1007 22:21:35.513496 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:21:35 crc kubenswrapper[4871]: I1007 22:21:35.514436 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f191ac46661d036cac4f07f7277192def3c4b3a3a733b08403472441f085e8e0"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:21:35 crc kubenswrapper[4871]: I1007 22:21:35.514513 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://f191ac46661d036cac4f07f7277192def3c4b3a3a733b08403472441f085e8e0" gracePeriod=600 Oct 07 22:21:36 crc kubenswrapper[4871]: I1007 22:21:36.340725 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="f191ac46661d036cac4f07f7277192def3c4b3a3a733b08403472441f085e8e0" exitCode=0 Oct 07 22:21:36 crc kubenswrapper[4871]: I1007 22:21:36.340860 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"f191ac46661d036cac4f07f7277192def3c4b3a3a733b08403472441f085e8e0"} Oct 07 22:21:36 crc kubenswrapper[4871]: I1007 22:21:36.341685 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"5fbc94adc3f0888dbb1bfd5173d9407ea8c5c758052b4cd3dbe78d9b527973c4"} Oct 07 22:21:36 crc kubenswrapper[4871]: I1007 22:21:36.341722 4871 scope.go:117] "RemoveContainer" containerID="fe9062423d89d34c8005bfab202352c48e57c724729bdae8f94451cff70b4074" Oct 07 22:21:39 crc kubenswrapper[4871]: I1007 22:21:39.981712 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:39 crc kubenswrapper[4871]: I1007 22:21:39.983198 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:40 crc kubenswrapper[4871]: I1007 22:21:40.510671 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-hntcc"] Oct 07 22:21:40 crc kubenswrapper[4871]: W1007 22:21:40.522250 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7cd42c4_268b_4588_a427_0a80502fe808.slice/crio-a5b0b51e6688c0050b68763347d0db105bbbe999aee812adfcd1ae9b52bec280 WatchSource:0}: Error finding container a5b0b51e6688c0050b68763347d0db105bbbe999aee812adfcd1ae9b52bec280: Status 404 returned error can't find the container with id a5b0b51e6688c0050b68763347d0db105bbbe999aee812adfcd1ae9b52bec280 Oct 07 22:21:40 crc kubenswrapper[4871]: I1007 22:21:40.525759 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:21:40 crc kubenswrapper[4871]: I1007 22:21:40.983707 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:40 crc kubenswrapper[4871]: I1007 22:21:40.985184 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:41 crc kubenswrapper[4871]: I1007 22:21:41.390509 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-hntcc" event={"ID":"d7cd42c4-268b-4588-a427-0a80502fe808","Type":"ContainerStarted","Data":"a5b0b51e6688c0050b68763347d0db105bbbe999aee812adfcd1ae9b52bec280"} Oct 07 22:21:41 crc kubenswrapper[4871]: I1007 22:21:41.518864 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q47sb"] Oct 07 22:21:41 crc kubenswrapper[4871]: W1007 22:21:41.525136 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6003e812_ee55_4cf6_b265_fe3f242d7fab.slice/crio-2f927e07ce90769aef39b658cc6cc4db6fbd139d4d04afcd690b62680472329e WatchSource:0}: Error finding container 2f927e07ce90769aef39b658cc6cc4db6fbd139d4d04afcd690b62680472329e: Status 404 returned error can't find the container with id 2f927e07ce90769aef39b658cc6cc4db6fbd139d4d04afcd690b62680472329e Oct 07 22:21:42 crc kubenswrapper[4871]: I1007 22:21:42.400749 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q47sb" event={"ID":"6003e812-ee55-4cf6-b265-fe3f242d7fab","Type":"ContainerDied","Data":"aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9"} Oct 07 22:21:42 crc kubenswrapper[4871]: I1007 22:21:42.400690 4871 generic.go:334] "Generic (PLEG): container finished" podID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerID="aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9" exitCode=0 Oct 07 22:21:42 crc kubenswrapper[4871]: I1007 22:21:42.401315 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q47sb" event={"ID":"6003e812-ee55-4cf6-b265-fe3f242d7fab","Type":"ContainerStarted","Data":"2f927e07ce90769aef39b658cc6cc4db6fbd139d4d04afcd690b62680472329e"} Oct 07 22:21:43 crc kubenswrapper[4871]: I1007 22:21:43.412458 4871 generic.go:334] "Generic (PLEG): container finished" podID="d7cd42c4-268b-4588-a427-0a80502fe808" containerID="b373140d1a8bd7e0a2a7c1dab5e64d71766cf61ca69697fd4a99ad3ff18304f5" exitCode=0 Oct 07 22:21:43 crc kubenswrapper[4871]: I1007 22:21:43.412553 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-hntcc" event={"ID":"d7cd42c4-268b-4588-a427-0a80502fe808","Type":"ContainerDied","Data":"b373140d1a8bd7e0a2a7c1dab5e64d71766cf61ca69697fd4a99ad3ff18304f5"} Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.769125 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.878057 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d7cd42c4-268b-4588-a427-0a80502fe808-node-mnt\") pod \"d7cd42c4-268b-4588-a427-0a80502fe808\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.878104 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbbvm\" (UniqueName: \"kubernetes.io/projected/d7cd42c4-268b-4588-a427-0a80502fe808-kube-api-access-hbbvm\") pod \"d7cd42c4-268b-4588-a427-0a80502fe808\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.878148 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d7cd42c4-268b-4588-a427-0a80502fe808-crc-storage\") pod \"d7cd42c4-268b-4588-a427-0a80502fe808\" (UID: \"d7cd42c4-268b-4588-a427-0a80502fe808\") " Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.878179 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7cd42c4-268b-4588-a427-0a80502fe808-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "d7cd42c4-268b-4588-a427-0a80502fe808" (UID: "d7cd42c4-268b-4588-a427-0a80502fe808"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.878350 4871 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d7cd42c4-268b-4588-a427-0a80502fe808-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.884495 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7cd42c4-268b-4588-a427-0a80502fe808-kube-api-access-hbbvm" (OuterVolumeSpecName: "kube-api-access-hbbvm") pod "d7cd42c4-268b-4588-a427-0a80502fe808" (UID: "d7cd42c4-268b-4588-a427-0a80502fe808"). InnerVolumeSpecName "kube-api-access-hbbvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.897608 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7cd42c4-268b-4588-a427-0a80502fe808-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "d7cd42c4-268b-4588-a427-0a80502fe808" (UID: "d7cd42c4-268b-4588-a427-0a80502fe808"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.979257 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbbvm\" (UniqueName: \"kubernetes.io/projected/d7cd42c4-268b-4588-a427-0a80502fe808-kube-api-access-hbbvm\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:44 crc kubenswrapper[4871]: I1007 22:21:44.979300 4871 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d7cd42c4-268b-4588-a427-0a80502fe808-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.407873 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-72npj"] Oct 07 22:21:45 crc kubenswrapper[4871]: E1007 22:21:45.408158 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7cd42c4-268b-4588-a427-0a80502fe808" containerName="storage" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.408176 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7cd42c4-268b-4588-a427-0a80502fe808" containerName="storage" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.408331 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7cd42c4-268b-4588-a427-0a80502fe808" containerName="storage" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.410579 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.424753 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72npj"] Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.436154 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-hntcc" event={"ID":"d7cd42c4-268b-4588-a427-0a80502fe808","Type":"ContainerDied","Data":"a5b0b51e6688c0050b68763347d0db105bbbe999aee812adfcd1ae9b52bec280"} Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.436213 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5b0b51e6688c0050b68763347d0db105bbbe999aee812adfcd1ae9b52bec280" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.436337 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hntcc" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.441058 4871 generic.go:334] "Generic (PLEG): container finished" podID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerID="1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624" exitCode=0 Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.441123 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q47sb" event={"ID":"6003e812-ee55-4cf6-b265-fe3f242d7fab","Type":"ContainerDied","Data":"1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624"} Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.489611 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-utilities\") pod \"community-operators-72npj\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.489873 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-catalog-content\") pod \"community-operators-72npj\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.490030 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbfpd\" (UniqueName: \"kubernetes.io/projected/79a81358-961f-4983-b529-fa2cc08aa95c-kube-api-access-mbfpd\") pod \"community-operators-72npj\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.591570 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-utilities\") pod \"community-operators-72npj\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.591643 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-catalog-content\") pod \"community-operators-72npj\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.591684 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbfpd\" (UniqueName: \"kubernetes.io/projected/79a81358-961f-4983-b529-fa2cc08aa95c-kube-api-access-mbfpd\") pod \"community-operators-72npj\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.592259 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-catalog-content\") pod \"community-operators-72npj\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.592299 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-utilities\") pod \"community-operators-72npj\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.613327 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbfpd\" (UniqueName: \"kubernetes.io/projected/79a81358-961f-4983-b529-fa2cc08aa95c-kube-api-access-mbfpd\") pod \"community-operators-72npj\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:45 crc kubenswrapper[4871]: I1007 22:21:45.746220 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:46 crc kubenswrapper[4871]: I1007 22:21:46.300146 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72npj"] Oct 07 22:21:46 crc kubenswrapper[4871]: I1007 22:21:46.450383 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72npj" event={"ID":"79a81358-961f-4983-b529-fa2cc08aa95c","Type":"ContainerStarted","Data":"2011804d9a47684be94e08cbf60f70da4a3a206519421cb43d907e270a0779b4"} Oct 07 22:21:47 crc kubenswrapper[4871]: I1007 22:21:47.461527 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q47sb" event={"ID":"6003e812-ee55-4cf6-b265-fe3f242d7fab","Type":"ContainerStarted","Data":"5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70"} Oct 07 22:21:47 crc kubenswrapper[4871]: I1007 22:21:47.466739 4871 generic.go:334] "Generic (PLEG): container finished" podID="79a81358-961f-4983-b529-fa2cc08aa95c" containerID="cd8794e5f47452407dd1e8f41bc661b06b823453d8a04120b44823036660bc90" exitCode=0 Oct 07 22:21:47 crc kubenswrapper[4871]: I1007 22:21:47.466988 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72npj" event={"ID":"79a81358-961f-4983-b529-fa2cc08aa95c","Type":"ContainerDied","Data":"cd8794e5f47452407dd1e8f41bc661b06b823453d8a04120b44823036660bc90"} Oct 07 22:21:47 crc kubenswrapper[4871]: I1007 22:21:47.490085 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q47sb" podStartSLOduration=15.435758746 podStartE2EDuration="19.490063592s" podCreationTimestamp="2025-10-07 22:21:28 +0000 UTC" firstStartedPulling="2025-10-07 22:21:42.496203705 +0000 UTC m=+776.298901778" lastFinishedPulling="2025-10-07 22:21:46.550508551 +0000 UTC m=+780.353206624" observedRunningTime="2025-10-07 22:21:47.486675543 +0000 UTC m=+781.289373616" watchObservedRunningTime="2025-10-07 22:21:47.490063592 +0000 UTC m=+781.292761665" Oct 07 22:21:49 crc kubenswrapper[4871]: I1007 22:21:49.049366 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:49 crc kubenswrapper[4871]: I1007 22:21:49.049958 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:49 crc kubenswrapper[4871]: I1007 22:21:49.115509 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:50 crc kubenswrapper[4871]: I1007 22:21:50.201202 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8h4fg" Oct 07 22:21:50 crc kubenswrapper[4871]: I1007 22:21:50.491802 4871 generic.go:334] "Generic (PLEG): container finished" podID="79a81358-961f-4983-b529-fa2cc08aa95c" containerID="e99f17b884e1d5f537ffa65b85074b4946166e47c33cd5a19021116909beca98" exitCode=0 Oct 07 22:21:50 crc kubenswrapper[4871]: I1007 22:21:50.492018 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72npj" event={"ID":"79a81358-961f-4983-b529-fa2cc08aa95c","Type":"ContainerDied","Data":"e99f17b884e1d5f537ffa65b85074b4946166e47c33cd5a19021116909beca98"} Oct 07 22:21:52 crc kubenswrapper[4871]: I1007 22:21:52.506115 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72npj" event={"ID":"79a81358-961f-4983-b529-fa2cc08aa95c","Type":"ContainerStarted","Data":"6f440e90566f9d0756ae60a1cc9d22debdd1879a39aaf1b73bff22b184785357"} Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.045478 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-72npj" podStartSLOduration=3.724536391 podStartE2EDuration="8.045451971s" podCreationTimestamp="2025-10-07 22:21:45 +0000 UTC" firstStartedPulling="2025-10-07 22:21:47.468469584 +0000 UTC m=+781.271167657" lastFinishedPulling="2025-10-07 22:21:51.789385164 +0000 UTC m=+785.592083237" observedRunningTime="2025-10-07 22:21:52.52910488 +0000 UTC m=+786.331802953" watchObservedRunningTime="2025-10-07 22:21:53.045451971 +0000 UTC m=+786.848150074" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.047012 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m"] Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.048784 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.052671 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.067639 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m"] Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.125781 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.125871 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.125945 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzvtd\" (UniqueName: \"kubernetes.io/projected/a34d9156-1901-41f3-a16c-c80d992569bb-kube-api-access-zzvtd\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.227951 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.228043 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.228164 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzvtd\" (UniqueName: \"kubernetes.io/projected/a34d9156-1901-41f3-a16c-c80d992569bb-kube-api-access-zzvtd\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.228751 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.229154 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.263685 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzvtd\" (UniqueName: \"kubernetes.io/projected/a34d9156-1901-41f3-a16c-c80d992569bb-kube-api-access-zzvtd\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.378155 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:21:53 crc kubenswrapper[4871]: I1007 22:21:53.905894 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m"] Oct 07 22:21:53 crc kubenswrapper[4871]: W1007 22:21:53.919904 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda34d9156_1901_41f3_a16c_c80d992569bb.slice/crio-d0919975bf1344474b5630a17962308d2a2dd3e107308bfada913a58ac683ea3 WatchSource:0}: Error finding container d0919975bf1344474b5630a17962308d2a2dd3e107308bfada913a58ac683ea3: Status 404 returned error can't find the container with id d0919975bf1344474b5630a17962308d2a2dd3e107308bfada913a58ac683ea3 Oct 07 22:21:54 crc kubenswrapper[4871]: I1007 22:21:54.522432 4871 generic.go:334] "Generic (PLEG): container finished" podID="a34d9156-1901-41f3-a16c-c80d992569bb" containerID="709aaca48827f50c4645788e3eebc4c17ed25a80e19abca92cc88086cb156786" exitCode=0 Oct 07 22:21:54 crc kubenswrapper[4871]: I1007 22:21:54.522528 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" event={"ID":"a34d9156-1901-41f3-a16c-c80d992569bb","Type":"ContainerDied","Data":"709aaca48827f50c4645788e3eebc4c17ed25a80e19abca92cc88086cb156786"} Oct 07 22:21:54 crc kubenswrapper[4871]: I1007 22:21:54.522970 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" event={"ID":"a34d9156-1901-41f3-a16c-c80d992569bb","Type":"ContainerStarted","Data":"d0919975bf1344474b5630a17962308d2a2dd3e107308bfada913a58ac683ea3"} Oct 07 22:21:55 crc kubenswrapper[4871]: I1007 22:21:55.747270 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:55 crc kubenswrapper[4871]: I1007 22:21:55.747772 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:55 crc kubenswrapper[4871]: I1007 22:21:55.812569 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.182628 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bzltp"] Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.183661 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.204944 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bzltp"] Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.280625 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbwkn\" (UniqueName: \"kubernetes.io/projected/e7bf53ee-7427-43df-9813-d059e6765fe7-kube-api-access-pbwkn\") pod \"redhat-operators-bzltp\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.280686 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-catalog-content\") pod \"redhat-operators-bzltp\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.281034 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-utilities\") pod \"redhat-operators-bzltp\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.382978 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbwkn\" (UniqueName: \"kubernetes.io/projected/e7bf53ee-7427-43df-9813-d059e6765fe7-kube-api-access-pbwkn\") pod \"redhat-operators-bzltp\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.383034 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-catalog-content\") pod \"redhat-operators-bzltp\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.383075 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-utilities\") pod \"redhat-operators-bzltp\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.383624 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-utilities\") pod \"redhat-operators-bzltp\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.383705 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-catalog-content\") pod \"redhat-operators-bzltp\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.405079 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbwkn\" (UniqueName: \"kubernetes.io/projected/e7bf53ee-7427-43df-9813-d059e6765fe7-kube-api-access-pbwkn\") pod \"redhat-operators-bzltp\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.519216 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.581966 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:56 crc kubenswrapper[4871]: I1007 22:21:56.966274 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bzltp"] Oct 07 22:21:57 crc kubenswrapper[4871]: I1007 22:21:57.546542 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" event={"ID":"a34d9156-1901-41f3-a16c-c80d992569bb","Type":"ContainerStarted","Data":"278796487abcb588686e836b7a7af0d422fad95cb9ca8f344e286e712b4de258"} Oct 07 22:21:57 crc kubenswrapper[4871]: I1007 22:21:57.551441 4871 generic.go:334] "Generic (PLEG): container finished" podID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerID="a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6" exitCode=0 Oct 07 22:21:57 crc kubenswrapper[4871]: I1007 22:21:57.551553 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzltp" event={"ID":"e7bf53ee-7427-43df-9813-d059e6765fe7","Type":"ContainerDied","Data":"a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6"} Oct 07 22:21:57 crc kubenswrapper[4871]: I1007 22:21:57.551623 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzltp" event={"ID":"e7bf53ee-7427-43df-9813-d059e6765fe7","Type":"ContainerStarted","Data":"1f2dca5943aff2e1f2324bede5b295d48051fdc3517c8d844e54dd3ca291e2c7"} Oct 07 22:21:58 crc kubenswrapper[4871]: I1007 22:21:58.564606 4871 generic.go:334] "Generic (PLEG): container finished" podID="a34d9156-1901-41f3-a16c-c80d992569bb" containerID="278796487abcb588686e836b7a7af0d422fad95cb9ca8f344e286e712b4de258" exitCode=0 Oct 07 22:21:58 crc kubenswrapper[4871]: I1007 22:21:58.564692 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" event={"ID":"a34d9156-1901-41f3-a16c-c80d992569bb","Type":"ContainerDied","Data":"278796487abcb588686e836b7a7af0d422fad95cb9ca8f344e286e712b4de258"} Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.113413 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.369626 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72npj"] Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.369970 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-72npj" podUID="79a81358-961f-4983-b529-fa2cc08aa95c" containerName="registry-server" containerID="cri-o://6f440e90566f9d0756ae60a1cc9d22debdd1879a39aaf1b73bff22b184785357" gracePeriod=2 Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.576559 4871 generic.go:334] "Generic (PLEG): container finished" podID="a34d9156-1901-41f3-a16c-c80d992569bb" containerID="21985f8943df30b77d2bb188b689ea4f156363d865bba73619141761272654f1" exitCode=0 Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.576675 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" event={"ID":"a34d9156-1901-41f3-a16c-c80d992569bb","Type":"ContainerDied","Data":"21985f8943df30b77d2bb188b689ea4f156363d865bba73619141761272654f1"} Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.580237 4871 generic.go:334] "Generic (PLEG): container finished" podID="79a81358-961f-4983-b529-fa2cc08aa95c" containerID="6f440e90566f9d0756ae60a1cc9d22debdd1879a39aaf1b73bff22b184785357" exitCode=0 Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.580320 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72npj" event={"ID":"79a81358-961f-4983-b529-fa2cc08aa95c","Type":"ContainerDied","Data":"6f440e90566f9d0756ae60a1cc9d22debdd1879a39aaf1b73bff22b184785357"} Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.858576 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72npj" Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.942605 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbfpd\" (UniqueName: \"kubernetes.io/projected/79a81358-961f-4983-b529-fa2cc08aa95c-kube-api-access-mbfpd\") pod \"79a81358-961f-4983-b529-fa2cc08aa95c\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.942699 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-utilities\") pod \"79a81358-961f-4983-b529-fa2cc08aa95c\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.942751 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-catalog-content\") pod \"79a81358-961f-4983-b529-fa2cc08aa95c\" (UID: \"79a81358-961f-4983-b529-fa2cc08aa95c\") " Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.944201 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-utilities" (OuterVolumeSpecName: "utilities") pod "79a81358-961f-4983-b529-fa2cc08aa95c" (UID: "79a81358-961f-4983-b529-fa2cc08aa95c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:21:59 crc kubenswrapper[4871]: I1007 22:21:59.953155 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a81358-961f-4983-b529-fa2cc08aa95c-kube-api-access-mbfpd" (OuterVolumeSpecName: "kube-api-access-mbfpd") pod "79a81358-961f-4983-b529-fa2cc08aa95c" (UID: "79a81358-961f-4983-b529-fa2cc08aa95c"). InnerVolumeSpecName "kube-api-access-mbfpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.017631 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79a81358-961f-4983-b529-fa2cc08aa95c" (UID: "79a81358-961f-4983-b529-fa2cc08aa95c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.044690 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbfpd\" (UniqueName: \"kubernetes.io/projected/79a81358-961f-4983-b529-fa2cc08aa95c-kube-api-access-mbfpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.044718 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.044729 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79a81358-961f-4983-b529-fa2cc08aa95c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.588847 4871 generic.go:334] "Generic (PLEG): container finished" podID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerID="6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457" exitCode=0 Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.588921 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzltp" event={"ID":"e7bf53ee-7427-43df-9813-d059e6765fe7","Type":"ContainerDied","Data":"6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457"} Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.596988 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72npj" event={"ID":"79a81358-961f-4983-b529-fa2cc08aa95c","Type":"ContainerDied","Data":"2011804d9a47684be94e08cbf60f70da4a3a206519421cb43d907e270a0779b4"} Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.597071 4871 scope.go:117] "RemoveContainer" containerID="6f440e90566f9d0756ae60a1cc9d22debdd1879a39aaf1b73bff22b184785357" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.597262 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72npj" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.628002 4871 scope.go:117] "RemoveContainer" containerID="e99f17b884e1d5f537ffa65b85074b4946166e47c33cd5a19021116909beca98" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.650833 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72npj"] Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.655847 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-72npj"] Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.670130 4871 scope.go:117] "RemoveContainer" containerID="cd8794e5f47452407dd1e8f41bc661b06b823453d8a04120b44823036660bc90" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.984637 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:22:00 crc kubenswrapper[4871]: I1007 22:22:00.995482 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a81358-961f-4983-b529-fa2cc08aa95c" path="/var/lib/kubelet/pods/79a81358-961f-4983-b529-fa2cc08aa95c/volumes" Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.056536 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzvtd\" (UniqueName: \"kubernetes.io/projected/a34d9156-1901-41f3-a16c-c80d992569bb-kube-api-access-zzvtd\") pod \"a34d9156-1901-41f3-a16c-c80d992569bb\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.056619 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-util\") pod \"a34d9156-1901-41f3-a16c-c80d992569bb\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.056763 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-bundle\") pod \"a34d9156-1901-41f3-a16c-c80d992569bb\" (UID: \"a34d9156-1901-41f3-a16c-c80d992569bb\") " Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.059105 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-bundle" (OuterVolumeSpecName: "bundle") pod "a34d9156-1901-41f3-a16c-c80d992569bb" (UID: "a34d9156-1901-41f3-a16c-c80d992569bb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.065242 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a34d9156-1901-41f3-a16c-c80d992569bb-kube-api-access-zzvtd" (OuterVolumeSpecName: "kube-api-access-zzvtd") pod "a34d9156-1901-41f3-a16c-c80d992569bb" (UID: "a34d9156-1901-41f3-a16c-c80d992569bb"). InnerVolumeSpecName "kube-api-access-zzvtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.072254 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-util" (OuterVolumeSpecName: "util") pod "a34d9156-1901-41f3-a16c-c80d992569bb" (UID: "a34d9156-1901-41f3-a16c-c80d992569bb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.159758 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzvtd\" (UniqueName: \"kubernetes.io/projected/a34d9156-1901-41f3-a16c-c80d992569bb-kube-api-access-zzvtd\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.159832 4871 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-util\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.159849 4871 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a34d9156-1901-41f3-a16c-c80d992569bb-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.607888 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.609144 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m" event={"ID":"a34d9156-1901-41f3-a16c-c80d992569bb","Type":"ContainerDied","Data":"d0919975bf1344474b5630a17962308d2a2dd3e107308bfada913a58ac683ea3"} Oct 07 22:22:01 crc kubenswrapper[4871]: I1007 22:22:01.609208 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0919975bf1344474b5630a17962308d2a2dd3e107308bfada913a58ac683ea3" Oct 07 22:22:02 crc kubenswrapper[4871]: I1007 22:22:02.622544 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzltp" event={"ID":"e7bf53ee-7427-43df-9813-d059e6765fe7","Type":"ContainerStarted","Data":"9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851"} Oct 07 22:22:02 crc kubenswrapper[4871]: I1007 22:22:02.655085 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bzltp" podStartSLOduration=2.405616196 podStartE2EDuration="6.655045705s" podCreationTimestamp="2025-10-07 22:21:56 +0000 UTC" firstStartedPulling="2025-10-07 22:21:57.552652781 +0000 UTC m=+791.355350894" lastFinishedPulling="2025-10-07 22:22:01.80208233 +0000 UTC m=+795.604780403" observedRunningTime="2025-10-07 22:22:02.649132569 +0000 UTC m=+796.451830682" watchObservedRunningTime="2025-10-07 22:22:02.655045705 +0000 UTC m=+796.457743818" Oct 07 22:22:02 crc kubenswrapper[4871]: I1007 22:22:02.770952 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q47sb"] Oct 07 22:22:02 crc kubenswrapper[4871]: I1007 22:22:02.771475 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q47sb" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerName="registry-server" containerID="cri-o://5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70" gracePeriod=2 Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.273590 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.392747 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kkkw\" (UniqueName: \"kubernetes.io/projected/6003e812-ee55-4cf6-b265-fe3f242d7fab-kube-api-access-9kkkw\") pod \"6003e812-ee55-4cf6-b265-fe3f242d7fab\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.393237 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-utilities\") pod \"6003e812-ee55-4cf6-b265-fe3f242d7fab\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.393317 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-catalog-content\") pod \"6003e812-ee55-4cf6-b265-fe3f242d7fab\" (UID: \"6003e812-ee55-4cf6-b265-fe3f242d7fab\") " Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.394140 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-utilities" (OuterVolumeSpecName: "utilities") pod "6003e812-ee55-4cf6-b265-fe3f242d7fab" (UID: "6003e812-ee55-4cf6-b265-fe3f242d7fab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.400120 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6003e812-ee55-4cf6-b265-fe3f242d7fab-kube-api-access-9kkkw" (OuterVolumeSpecName: "kube-api-access-9kkkw") pod "6003e812-ee55-4cf6-b265-fe3f242d7fab" (UID: "6003e812-ee55-4cf6-b265-fe3f242d7fab"). InnerVolumeSpecName "kube-api-access-9kkkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.407135 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6003e812-ee55-4cf6-b265-fe3f242d7fab" (UID: "6003e812-ee55-4cf6-b265-fe3f242d7fab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.494976 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.495028 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kkkw\" (UniqueName: \"kubernetes.io/projected/6003e812-ee55-4cf6-b265-fe3f242d7fab-kube-api-access-9kkkw\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.495046 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6003e812-ee55-4cf6-b265-fe3f242d7fab-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591045 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx"] Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.591374 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerName="extract-utilities" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591400 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerName="extract-utilities" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.591417 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34d9156-1901-41f3-a16c-c80d992569bb" containerName="pull" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591427 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34d9156-1901-41f3-a16c-c80d992569bb" containerName="pull" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.591445 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a81358-961f-4983-b529-fa2cc08aa95c" containerName="registry-server" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591454 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a81358-961f-4983-b529-fa2cc08aa95c" containerName="registry-server" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.591466 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerName="registry-server" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591478 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerName="registry-server" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.591497 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a81358-961f-4983-b529-fa2cc08aa95c" containerName="extract-utilities" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591506 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a81358-961f-4983-b529-fa2cc08aa95c" containerName="extract-utilities" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.591520 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a81358-961f-4983-b529-fa2cc08aa95c" containerName="extract-content" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591528 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a81358-961f-4983-b529-fa2cc08aa95c" containerName="extract-content" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.591536 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34d9156-1901-41f3-a16c-c80d992569bb" containerName="extract" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591544 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34d9156-1901-41f3-a16c-c80d992569bb" containerName="extract" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.591560 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34d9156-1901-41f3-a16c-c80d992569bb" containerName="util" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591567 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34d9156-1901-41f3-a16c-c80d992569bb" containerName="util" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.591578 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerName="extract-content" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591588 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerName="extract-content" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591711 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a34d9156-1901-41f3-a16c-c80d992569bb" containerName="extract" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591728 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerName="registry-server" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.591741 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a81358-961f-4983-b529-fa2cc08aa95c" containerName="registry-server" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.592365 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.594196 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.594926 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-xd2xh" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.595093 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.603245 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx"] Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.636957 4871 generic.go:334] "Generic (PLEG): container finished" podID="6003e812-ee55-4cf6-b265-fe3f242d7fab" containerID="5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70" exitCode=0 Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.637843 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q47sb" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.638020 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q47sb" event={"ID":"6003e812-ee55-4cf6-b265-fe3f242d7fab","Type":"ContainerDied","Data":"5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70"} Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.638056 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q47sb" event={"ID":"6003e812-ee55-4cf6-b265-fe3f242d7fab","Type":"ContainerDied","Data":"2f927e07ce90769aef39b658cc6cc4db6fbd139d4d04afcd690b62680472329e"} Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.638081 4871 scope.go:117] "RemoveContainer" containerID="5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.657777 4871 scope.go:117] "RemoveContainer" containerID="1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.669876 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q47sb"] Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.682571 4871 scope.go:117] "RemoveContainer" containerID="aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.686264 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q47sb"] Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.696991 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcsb9\" (UniqueName: \"kubernetes.io/projected/8c4cee14-12e0-4ce6-9141-cfa251152a35-kube-api-access-tcsb9\") pod \"nmstate-operator-858ddd8f98-lvfxx\" (UID: \"8c4cee14-12e0-4ce6-9141-cfa251152a35\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.702720 4871 scope.go:117] "RemoveContainer" containerID="5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.704134 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70\": container with ID starting with 5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70 not found: ID does not exist" containerID="5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.704196 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70"} err="failed to get container status \"5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70\": rpc error: code = NotFound desc = could not find container \"5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70\": container with ID starting with 5feb94182a10079d4b3d5c4dcd501cd74ef5cf7607740d14899451bc890e9c70 not found: ID does not exist" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.704229 4871 scope.go:117] "RemoveContainer" containerID="1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.704722 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624\": container with ID starting with 1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624 not found: ID does not exist" containerID="1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.704915 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624"} err="failed to get container status \"1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624\": rpc error: code = NotFound desc = could not find container \"1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624\": container with ID starting with 1a393b85c51ab54e051033e6a3f4b32d6e50df08e3dae0bc4cc626fd6e315624 not found: ID does not exist" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.705019 4871 scope.go:117] "RemoveContainer" containerID="aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9" Oct 07 22:22:03 crc kubenswrapper[4871]: E1007 22:22:03.705606 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9\": container with ID starting with aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9 not found: ID does not exist" containerID="aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.705661 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9"} err="failed to get container status \"aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9\": rpc error: code = NotFound desc = could not find container \"aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9\": container with ID starting with aebce3036eaba3edd1a80c3ad53f2db447c3ae56cd793056aac23e7adf8e49f9 not found: ID does not exist" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.799012 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcsb9\" (UniqueName: \"kubernetes.io/projected/8c4cee14-12e0-4ce6-9141-cfa251152a35-kube-api-access-tcsb9\") pod \"nmstate-operator-858ddd8f98-lvfxx\" (UID: \"8c4cee14-12e0-4ce6-9141-cfa251152a35\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.816398 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcsb9\" (UniqueName: \"kubernetes.io/projected/8c4cee14-12e0-4ce6-9141-cfa251152a35-kube-api-access-tcsb9\") pod \"nmstate-operator-858ddd8f98-lvfxx\" (UID: \"8c4cee14-12e0-4ce6-9141-cfa251152a35\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx" Oct 07 22:22:03 crc kubenswrapper[4871]: I1007 22:22:03.911987 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx" Oct 07 22:22:04 crc kubenswrapper[4871]: I1007 22:22:04.357336 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx"] Oct 07 22:22:04 crc kubenswrapper[4871]: I1007 22:22:04.644133 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx" event={"ID":"8c4cee14-12e0-4ce6-9141-cfa251152a35","Type":"ContainerStarted","Data":"29fb4277d0c0e215151291d3a784d6c15992542aef0dddea505fda8e6ce4ab90"} Oct 07 22:22:04 crc kubenswrapper[4871]: I1007 22:22:04.989609 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6003e812-ee55-4cf6-b265-fe3f242d7fab" path="/var/lib/kubelet/pods/6003e812-ee55-4cf6-b265-fe3f242d7fab/volumes" Oct 07 22:22:06 crc kubenswrapper[4871]: I1007 22:22:06.520220 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:22:06 crc kubenswrapper[4871]: I1007 22:22:06.520624 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:22:07 crc kubenswrapper[4871]: I1007 22:22:07.567822 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bzltp" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerName="registry-server" probeResult="failure" output=< Oct 07 22:22:07 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 22:22:07 crc kubenswrapper[4871]: > Oct 07 22:22:09 crc kubenswrapper[4871]: I1007 22:22:09.684386 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx" event={"ID":"8c4cee14-12e0-4ce6-9141-cfa251152a35","Type":"ContainerStarted","Data":"e70246ffd6c3082fe1dc700eca9a8df4ed836f4feb5f934de0d73203270febfb"} Oct 07 22:22:09 crc kubenswrapper[4871]: I1007 22:22:09.701353 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lvfxx" podStartSLOduration=2.23030327 podStartE2EDuration="6.701332758s" podCreationTimestamp="2025-10-07 22:22:03 +0000 UTC" firstStartedPulling="2025-10-07 22:22:04.36685501 +0000 UTC m=+798.169553083" lastFinishedPulling="2025-10-07 22:22:08.837884498 +0000 UTC m=+802.640582571" observedRunningTime="2025-10-07 22:22:09.698962496 +0000 UTC m=+803.501660569" watchObservedRunningTime="2025-10-07 22:22:09.701332758 +0000 UTC m=+803.504030821" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.095599 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.097051 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.100936 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-fpb85" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.107921 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.127346 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.135018 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-gvgbf"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.135295 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.136458 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.144893 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.189109 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.230966 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-dbus-socket\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.231013 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/11e73801-01a4-44c0-aaf8-5ad1e2af2239-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-v8gn8\" (UID: \"11e73801-01a4-44c0-aaf8-5ad1e2af2239\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.231062 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c6fc\" (UniqueName: \"kubernetes.io/projected/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-kube-api-access-5c6fc\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.231083 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sshtm\" (UniqueName: \"kubernetes.io/projected/058ddfce-22d3-45ea-8f8a-9463858b9d64-kube-api-access-sshtm\") pod \"nmstate-metrics-fdff9cb8d-hfbhn\" (UID: \"058ddfce-22d3-45ea-8f8a-9463858b9d64\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.231101 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-ovs-socket\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.231138 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-nmstate-lock\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.231164 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhszv\" (UniqueName: \"kubernetes.io/projected/11e73801-01a4-44c0-aaf8-5ad1e2af2239-kube-api-access-nhszv\") pod \"nmstate-webhook-6cdbc54649-v8gn8\" (UID: \"11e73801-01a4-44c0-aaf8-5ad1e2af2239\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.240554 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.241254 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.243246 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-nlr7s" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.243485 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.243897 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.258780 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.332171 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe9e41ea-e4e6-4c17-bd98-50ad8216b160-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-vs759\" (UID: \"fe9e41ea-e4e6-4c17-bd98-50ad8216b160\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.332517 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c6fc\" (UniqueName: \"kubernetes.io/projected/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-kube-api-access-5c6fc\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.332612 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sshtm\" (UniqueName: \"kubernetes.io/projected/058ddfce-22d3-45ea-8f8a-9463858b9d64-kube-api-access-sshtm\") pod \"nmstate-metrics-fdff9cb8d-hfbhn\" (UID: \"058ddfce-22d3-45ea-8f8a-9463858b9d64\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.332688 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-ovs-socket\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.332806 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-nmstate-lock\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.332867 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-nmstate-lock\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.332889 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgpkn\" (UniqueName: \"kubernetes.io/projected/fe9e41ea-e4e6-4c17-bd98-50ad8216b160-kube-api-access-xgpkn\") pod \"nmstate-console-plugin-6b874cbd85-vs759\" (UID: \"fe9e41ea-e4e6-4c17-bd98-50ad8216b160\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.332850 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-ovs-socket\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.333174 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhszv\" (UniqueName: \"kubernetes.io/projected/11e73801-01a4-44c0-aaf8-5ad1e2af2239-kube-api-access-nhszv\") pod \"nmstate-webhook-6cdbc54649-v8gn8\" (UID: \"11e73801-01a4-44c0-aaf8-5ad1e2af2239\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.333244 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-dbus-socket\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.333270 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/11e73801-01a4-44c0-aaf8-5ad1e2af2239-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-v8gn8\" (UID: \"11e73801-01a4-44c0-aaf8-5ad1e2af2239\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.333291 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fe9e41ea-e4e6-4c17-bd98-50ad8216b160-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-vs759\" (UID: \"fe9e41ea-e4e6-4c17-bd98-50ad8216b160\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: E1007 22:22:13.333589 4871 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 07 22:22:13 crc kubenswrapper[4871]: E1007 22:22:13.333662 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/11e73801-01a4-44c0-aaf8-5ad1e2af2239-tls-key-pair podName:11e73801-01a4-44c0-aaf8-5ad1e2af2239 nodeName:}" failed. No retries permitted until 2025-10-07 22:22:13.833632626 +0000 UTC m=+807.636330709 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/11e73801-01a4-44c0-aaf8-5ad1e2af2239-tls-key-pair") pod "nmstate-webhook-6cdbc54649-v8gn8" (UID: "11e73801-01a4-44c0-aaf8-5ad1e2af2239") : secret "openshift-nmstate-webhook" not found Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.333727 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-dbus-socket\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.354861 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c6fc\" (UniqueName: \"kubernetes.io/projected/b22d7c95-54d4-4fd5-87fa-3e6cf69a6932-kube-api-access-5c6fc\") pod \"nmstate-handler-gvgbf\" (UID: \"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932\") " pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.359842 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhszv\" (UniqueName: \"kubernetes.io/projected/11e73801-01a4-44c0-aaf8-5ad1e2af2239-kube-api-access-nhszv\") pod \"nmstate-webhook-6cdbc54649-v8gn8\" (UID: \"11e73801-01a4-44c0-aaf8-5ad1e2af2239\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.360182 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sshtm\" (UniqueName: \"kubernetes.io/projected/058ddfce-22d3-45ea-8f8a-9463858b9d64-kube-api-access-sshtm\") pod \"nmstate-metrics-fdff9cb8d-hfbhn\" (UID: \"058ddfce-22d3-45ea-8f8a-9463858b9d64\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.416970 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.437013 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe9e41ea-e4e6-4c17-bd98-50ad8216b160-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-vs759\" (UID: \"fe9e41ea-e4e6-4c17-bd98-50ad8216b160\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.437110 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgpkn\" (UniqueName: \"kubernetes.io/projected/fe9e41ea-e4e6-4c17-bd98-50ad8216b160-kube-api-access-xgpkn\") pod \"nmstate-console-plugin-6b874cbd85-vs759\" (UID: \"fe9e41ea-e4e6-4c17-bd98-50ad8216b160\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.437187 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fe9e41ea-e4e6-4c17-bd98-50ad8216b160-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-vs759\" (UID: \"fe9e41ea-e4e6-4c17-bd98-50ad8216b160\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.438758 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fe9e41ea-e4e6-4c17-bd98-50ad8216b160-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-vs759\" (UID: \"fe9e41ea-e4e6-4c17-bd98-50ad8216b160\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.444731 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-664b9b86b7-ss9xw"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.445914 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fe9e41ea-e4e6-4c17-bd98-50ad8216b160-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-vs759\" (UID: \"fe9e41ea-e4e6-4c17-bd98-50ad8216b160\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.445989 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.461251 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgpkn\" (UniqueName: \"kubernetes.io/projected/fe9e41ea-e4e6-4c17-bd98-50ad8216b160-kube-api-access-xgpkn\") pod \"nmstate-console-plugin-6b874cbd85-vs759\" (UID: \"fe9e41ea-e4e6-4c17-bd98-50ad8216b160\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.478836 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-664b9b86b7-ss9xw"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.505971 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.538733 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-console-config\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.539163 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-console-oauth-config\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.539218 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-trusted-ca-bundle\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.539246 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-oauth-serving-cert\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.539277 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-console-serving-cert\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.539294 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-service-ca\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.539314 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzmtx\" (UniqueName: \"kubernetes.io/projected/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-kube-api-access-pzmtx\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: W1007 22:22:13.555173 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb22d7c95_54d4_4fd5_87fa_3e6cf69a6932.slice/crio-e7c4f6c1e19e76c5a051faa54581497569750abf4bd1b61015fa58954f120cc1 WatchSource:0}: Error finding container e7c4f6c1e19e76c5a051faa54581497569750abf4bd1b61015fa58954f120cc1: Status 404 returned error can't find the container with id e7c4f6c1e19e76c5a051faa54581497569750abf4bd1b61015fa58954f120cc1 Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.567660 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.641202 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-console-config\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.641276 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-console-oauth-config\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.641330 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-trusted-ca-bundle\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.641364 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-oauth-serving-cert\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.641391 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-console-serving-cert\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.641406 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-service-ca\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.641423 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzmtx\" (UniqueName: \"kubernetes.io/projected/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-kube-api-access-pzmtx\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.643264 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-service-ca\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.643660 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-console-config\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.644133 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-trusted-ca-bundle\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.644682 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-oauth-serving-cert\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.647023 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-console-oauth-config\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.649476 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-console-serving-cert\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.657242 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzmtx\" (UniqueName: \"kubernetes.io/projected/4eec2665-4904-4cb0-b5b6-a5c40e41f1c4-kube-api-access-pzmtx\") pod \"console-664b9b86b7-ss9xw\" (UID: \"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4\") " pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.713154 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gvgbf" event={"ID":"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932","Type":"ContainerStarted","Data":"e7c4f6c1e19e76c5a051faa54581497569750abf4bd1b61015fa58954f120cc1"} Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.822454 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.844937 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/11e73801-01a4-44c0-aaf8-5ad1e2af2239-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-v8gn8\" (UID: \"11e73801-01a4-44c0-aaf8-5ad1e2af2239\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.851160 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/11e73801-01a4-44c0-aaf8-5ad1e2af2239-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-v8gn8\" (UID: \"11e73801-01a4-44c0-aaf8-5ad1e2af2239\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.941741 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn"] Oct 07 22:22:13 crc kubenswrapper[4871]: I1007 22:22:13.989308 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759"] Oct 07 22:22:14 crc kubenswrapper[4871]: W1007 22:22:14.008328 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe9e41ea_e4e6_4c17_bd98_50ad8216b160.slice/crio-f003280459047820cbe9ecd515d08217591523b430d1cd29f7ca7f3c6e0f30f0 WatchSource:0}: Error finding container f003280459047820cbe9ecd515d08217591523b430d1cd29f7ca7f3c6e0f30f0: Status 404 returned error can't find the container with id f003280459047820cbe9ecd515d08217591523b430d1cd29f7ca7f3c6e0f30f0 Oct 07 22:22:14 crc kubenswrapper[4871]: I1007 22:22:14.070054 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-664b9b86b7-ss9xw"] Oct 07 22:22:14 crc kubenswrapper[4871]: W1007 22:22:14.077400 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4eec2665_4904_4cb0_b5b6_a5c40e41f1c4.slice/crio-6fc635bfead1d1e270003ce68af57882c4d66b82bd3da10d3fabd53cfa6bab9b WatchSource:0}: Error finding container 6fc635bfead1d1e270003ce68af57882c4d66b82bd3da10d3fabd53cfa6bab9b: Status 404 returned error can't find the container with id 6fc635bfead1d1e270003ce68af57882c4d66b82bd3da10d3fabd53cfa6bab9b Oct 07 22:22:14 crc kubenswrapper[4871]: I1007 22:22:14.095045 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:14 crc kubenswrapper[4871]: I1007 22:22:14.532439 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8"] Oct 07 22:22:14 crc kubenswrapper[4871]: W1007 22:22:14.541486 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11e73801_01a4_44c0_aaf8_5ad1e2af2239.slice/crio-22b205b9fc7fe3e939fd907a8930d64846d0f1dc5f570cc2620ca91ddd20a92d WatchSource:0}: Error finding container 22b205b9fc7fe3e939fd907a8930d64846d0f1dc5f570cc2620ca91ddd20a92d: Status 404 returned error can't find the container with id 22b205b9fc7fe3e939fd907a8930d64846d0f1dc5f570cc2620ca91ddd20a92d Oct 07 22:22:14 crc kubenswrapper[4871]: I1007 22:22:14.722527 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" event={"ID":"fe9e41ea-e4e6-4c17-bd98-50ad8216b160","Type":"ContainerStarted","Data":"f003280459047820cbe9ecd515d08217591523b430d1cd29f7ca7f3c6e0f30f0"} Oct 07 22:22:14 crc kubenswrapper[4871]: I1007 22:22:14.724413 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn" event={"ID":"058ddfce-22d3-45ea-8f8a-9463858b9d64","Type":"ContainerStarted","Data":"8f6bf7a57f408edc38c0862ba327e0b936a609f4ad84758faf5fabb8bc0b8323"} Oct 07 22:22:14 crc kubenswrapper[4871]: I1007 22:22:14.726028 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" event={"ID":"11e73801-01a4-44c0-aaf8-5ad1e2af2239","Type":"ContainerStarted","Data":"22b205b9fc7fe3e939fd907a8930d64846d0f1dc5f570cc2620ca91ddd20a92d"} Oct 07 22:22:14 crc kubenswrapper[4871]: I1007 22:22:14.727898 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-664b9b86b7-ss9xw" event={"ID":"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4","Type":"ContainerStarted","Data":"fc9f9baa8a0eecbebed4b8da611de98f30ec7bc34dcab86ab5727ad0fdceba2b"} Oct 07 22:22:14 crc kubenswrapper[4871]: I1007 22:22:14.727971 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-664b9b86b7-ss9xw" event={"ID":"4eec2665-4904-4cb0-b5b6-a5c40e41f1c4","Type":"ContainerStarted","Data":"6fc635bfead1d1e270003ce68af57882c4d66b82bd3da10d3fabd53cfa6bab9b"} Oct 07 22:22:14 crc kubenswrapper[4871]: I1007 22:22:14.754237 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-664b9b86b7-ss9xw" podStartSLOduration=1.754216021 podStartE2EDuration="1.754216021s" podCreationTimestamp="2025-10-07 22:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:22:14.752462325 +0000 UTC m=+808.555160418" watchObservedRunningTime="2025-10-07 22:22:14.754216021 +0000 UTC m=+808.556914094" Oct 07 22:22:16 crc kubenswrapper[4871]: I1007 22:22:16.581280 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:22:16 crc kubenswrapper[4871]: I1007 22:22:16.647348 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:22:17 crc kubenswrapper[4871]: I1007 22:22:17.570311 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bzltp"] Oct 07 22:22:17 crc kubenswrapper[4871]: I1007 22:22:17.751117 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bzltp" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerName="registry-server" containerID="cri-o://9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851" gracePeriod=2 Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.705655 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.760159 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" event={"ID":"11e73801-01a4-44c0-aaf8-5ad1e2af2239","Type":"ContainerStarted","Data":"f2d55bba54113c51b9789beeb1672b69c8dab8d88eb989f614f6713ad197f80c"} Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.760467 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.765386 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" event={"ID":"fe9e41ea-e4e6-4c17-bd98-50ad8216b160","Type":"ContainerStarted","Data":"24dc0466e8056d2856f75adf45b37d26af3f09ab593af2b530bd03156649b1fb"} Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.768674 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn" event={"ID":"058ddfce-22d3-45ea-8f8a-9463858b9d64","Type":"ContainerStarted","Data":"8ee7f37e00c73c04899d91ccafa4a8115c6299055a9c909ab8dc8e6e671ae092"} Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.770975 4871 generic.go:334] "Generic (PLEG): container finished" podID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerID="9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851" exitCode=0 Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.771009 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzltp" event={"ID":"e7bf53ee-7427-43df-9813-d059e6765fe7","Type":"ContainerDied","Data":"9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851"} Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.771027 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzltp" event={"ID":"e7bf53ee-7427-43df-9813-d059e6765fe7","Type":"ContainerDied","Data":"1f2dca5943aff2e1f2324bede5b295d48051fdc3517c8d844e54dd3ca291e2c7"} Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.771047 4871 scope.go:117] "RemoveContainer" containerID="9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.771171 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzltp" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.783171 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" podStartSLOduration=1.8833475480000001 podStartE2EDuration="5.78315159s" podCreationTimestamp="2025-10-07 22:22:13 +0000 UTC" firstStartedPulling="2025-10-07 22:22:14.544082164 +0000 UTC m=+808.346780247" lastFinishedPulling="2025-10-07 22:22:18.443886206 +0000 UTC m=+812.246584289" observedRunningTime="2025-10-07 22:22:18.778911928 +0000 UTC m=+812.581610021" watchObservedRunningTime="2025-10-07 22:22:18.78315159 +0000 UTC m=+812.585849663" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.793446 4871 scope.go:117] "RemoveContainer" containerID="6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.810399 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-vs759" podStartSLOduration=1.392178718 podStartE2EDuration="5.810365165s" podCreationTimestamp="2025-10-07 22:22:13 +0000 UTC" firstStartedPulling="2025-10-07 22:22:14.013648012 +0000 UTC m=+807.816346085" lastFinishedPulling="2025-10-07 22:22:18.431834439 +0000 UTC m=+812.234532532" observedRunningTime="2025-10-07 22:22:18.80672134 +0000 UTC m=+812.609419413" watchObservedRunningTime="2025-10-07 22:22:18.810365165 +0000 UTC m=+812.613063228" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.811485 4871 scope.go:117] "RemoveContainer" containerID="a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.831721 4871 scope.go:117] "RemoveContainer" containerID="9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851" Oct 07 22:22:18 crc kubenswrapper[4871]: E1007 22:22:18.832296 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851\": container with ID starting with 9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851 not found: ID does not exist" containerID="9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.832363 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851"} err="failed to get container status \"9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851\": rpc error: code = NotFound desc = could not find container \"9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851\": container with ID starting with 9c2233d509dfbb75e60c5069bca7c05b190e151168ac4454441f4c19fb7e2851 not found: ID does not exist" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.832405 4871 scope.go:117] "RemoveContainer" containerID="6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457" Oct 07 22:22:18 crc kubenswrapper[4871]: E1007 22:22:18.833051 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457\": container with ID starting with 6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457 not found: ID does not exist" containerID="6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.833090 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457"} err="failed to get container status \"6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457\": rpc error: code = NotFound desc = could not find container \"6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457\": container with ID starting with 6ad39807dbcf40c02538e917392167a9026cec0dc329a28cd6f0b4b348a91457 not found: ID does not exist" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.833125 4871 scope.go:117] "RemoveContainer" containerID="a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6" Oct 07 22:22:18 crc kubenswrapper[4871]: E1007 22:22:18.833563 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6\": container with ID starting with a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6 not found: ID does not exist" containerID="a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.833601 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6"} err="failed to get container status \"a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6\": rpc error: code = NotFound desc = could not find container \"a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6\": container with ID starting with a9b89a4dca2319d616ea04731180ac3eca884a3e1f3662ace977412944b07fc6 not found: ID does not exist" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.883324 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-catalog-content\") pod \"e7bf53ee-7427-43df-9813-d059e6765fe7\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.883927 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-utilities\") pod \"e7bf53ee-7427-43df-9813-d059e6765fe7\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.883966 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbwkn\" (UniqueName: \"kubernetes.io/projected/e7bf53ee-7427-43df-9813-d059e6765fe7-kube-api-access-pbwkn\") pod \"e7bf53ee-7427-43df-9813-d059e6765fe7\" (UID: \"e7bf53ee-7427-43df-9813-d059e6765fe7\") " Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.884942 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-utilities" (OuterVolumeSpecName: "utilities") pod "e7bf53ee-7427-43df-9813-d059e6765fe7" (UID: "e7bf53ee-7427-43df-9813-d059e6765fe7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.890375 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7bf53ee-7427-43df-9813-d059e6765fe7-kube-api-access-pbwkn" (OuterVolumeSpecName: "kube-api-access-pbwkn") pod "e7bf53ee-7427-43df-9813-d059e6765fe7" (UID: "e7bf53ee-7427-43df-9813-d059e6765fe7"). InnerVolumeSpecName "kube-api-access-pbwkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.986614 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.987039 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbwkn\" (UniqueName: \"kubernetes.io/projected/e7bf53ee-7427-43df-9813-d059e6765fe7-kube-api-access-pbwkn\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:18 crc kubenswrapper[4871]: I1007 22:22:18.990830 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7bf53ee-7427-43df-9813-d059e6765fe7" (UID: "e7bf53ee-7427-43df-9813-d059e6765fe7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:19 crc kubenswrapper[4871]: I1007 22:22:19.092468 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7bf53ee-7427-43df-9813-d059e6765fe7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:19 crc kubenswrapper[4871]: I1007 22:22:19.095954 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bzltp"] Oct 07 22:22:19 crc kubenswrapper[4871]: I1007 22:22:19.103587 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bzltp"] Oct 07 22:22:19 crc kubenswrapper[4871]: I1007 22:22:19.784055 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gvgbf" event={"ID":"b22d7c95-54d4-4fd5-87fa-3e6cf69a6932","Type":"ContainerStarted","Data":"fc722bfe51ee52a9478fc08b48864fae8fe7a06df774ed1e8a870b85e60e6241"} Oct 07 22:22:19 crc kubenswrapper[4871]: I1007 22:22:19.806852 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-gvgbf" podStartSLOduration=1.937699957 podStartE2EDuration="6.806825904s" podCreationTimestamp="2025-10-07 22:22:13 +0000 UTC" firstStartedPulling="2025-10-07 22:22:13.557773772 +0000 UTC m=+807.360471845" lastFinishedPulling="2025-10-07 22:22:18.426899729 +0000 UTC m=+812.229597792" observedRunningTime="2025-10-07 22:22:19.804446972 +0000 UTC m=+813.607145045" watchObservedRunningTime="2025-10-07 22:22:19.806825904 +0000 UTC m=+813.609523987" Oct 07 22:22:20 crc kubenswrapper[4871]: I1007 22:22:20.789900 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:20 crc kubenswrapper[4871]: I1007 22:22:20.996401 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" path="/var/lib/kubelet/pods/e7bf53ee-7427-43df-9813-d059e6765fe7/volumes" Oct 07 22:22:22 crc kubenswrapper[4871]: I1007 22:22:22.807813 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn" event={"ID":"058ddfce-22d3-45ea-8f8a-9463858b9d64","Type":"ContainerStarted","Data":"3157695b742a12aec5041c50357281c98ab65e413ec2bdaac87c195621f4d33a"} Oct 07 22:22:22 crc kubenswrapper[4871]: I1007 22:22:22.833536 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-hfbhn" podStartSLOduration=1.421641534 podStartE2EDuration="9.833501973s" podCreationTimestamp="2025-10-07 22:22:13 +0000 UTC" firstStartedPulling="2025-10-07 22:22:13.951627251 +0000 UTC m=+807.754325364" lastFinishedPulling="2025-10-07 22:22:22.36348772 +0000 UTC m=+816.166185803" observedRunningTime="2025-10-07 22:22:22.828644195 +0000 UTC m=+816.631342268" watchObservedRunningTime="2025-10-07 22:22:22.833501973 +0000 UTC m=+816.636200076" Oct 07 22:22:23 crc kubenswrapper[4871]: I1007 22:22:23.550938 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-gvgbf" Oct 07 22:22:23 crc kubenswrapper[4871]: I1007 22:22:23.822933 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:23 crc kubenswrapper[4871]: I1007 22:22:23.823026 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:23 crc kubenswrapper[4871]: I1007 22:22:23.830703 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:24 crc kubenswrapper[4871]: I1007 22:22:24.832087 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-664b9b86b7-ss9xw" Oct 07 22:22:24 crc kubenswrapper[4871]: I1007 22:22:24.912827 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-tw5sf"] Oct 07 22:22:34 crc kubenswrapper[4871]: I1007 22:22:34.105609 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v8gn8" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.528449 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xs7cg"] Oct 07 22:22:44 crc kubenswrapper[4871]: E1007 22:22:44.530398 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerName="extract-utilities" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.530424 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerName="extract-utilities" Oct 07 22:22:44 crc kubenswrapper[4871]: E1007 22:22:44.530453 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerName="registry-server" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.530468 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerName="registry-server" Oct 07 22:22:44 crc kubenswrapper[4871]: E1007 22:22:44.530487 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerName="extract-content" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.530499 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerName="extract-content" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.530703 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7bf53ee-7427-43df-9813-d059e6765fe7" containerName="registry-server" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.533096 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xs7cg"] Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.533405 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.654680 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-catalog-content\") pod \"certified-operators-xs7cg\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.654757 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-utilities\") pod \"certified-operators-xs7cg\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.654818 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgqcc\" (UniqueName: \"kubernetes.io/projected/df014e0f-2f9f-4064-b7a8-db411cbe10b8-kube-api-access-lgqcc\") pod \"certified-operators-xs7cg\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.756762 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgqcc\" (UniqueName: \"kubernetes.io/projected/df014e0f-2f9f-4064-b7a8-db411cbe10b8-kube-api-access-lgqcc\") pod \"certified-operators-xs7cg\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.756917 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-catalog-content\") pod \"certified-operators-xs7cg\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.756947 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-utilities\") pod \"certified-operators-xs7cg\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.757612 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-utilities\") pod \"certified-operators-xs7cg\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.758442 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-catalog-content\") pod \"certified-operators-xs7cg\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.788688 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgqcc\" (UniqueName: \"kubernetes.io/projected/df014e0f-2f9f-4064-b7a8-db411cbe10b8-kube-api-access-lgqcc\") pod \"certified-operators-xs7cg\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:44 crc kubenswrapper[4871]: I1007 22:22:44.866164 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:45 crc kubenswrapper[4871]: I1007 22:22:45.397684 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xs7cg"] Oct 07 22:22:46 crc kubenswrapper[4871]: I1007 22:22:46.012337 4871 generic.go:334] "Generic (PLEG): container finished" podID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerID="a46685f20fb40e1e880918c97820e9c2e3e3fcaf6724bc3652b77f5b62226f21" exitCode=0 Oct 07 22:22:46 crc kubenswrapper[4871]: I1007 22:22:46.012409 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xs7cg" event={"ID":"df014e0f-2f9f-4064-b7a8-db411cbe10b8","Type":"ContainerDied","Data":"a46685f20fb40e1e880918c97820e9c2e3e3fcaf6724bc3652b77f5b62226f21"} Oct 07 22:22:46 crc kubenswrapper[4871]: I1007 22:22:46.012893 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xs7cg" event={"ID":"df014e0f-2f9f-4064-b7a8-db411cbe10b8","Type":"ContainerStarted","Data":"c76ab46b55d670d67ce7546132571864e8b549db2b07812801e6902e0219cd44"} Oct 07 22:22:48 crc kubenswrapper[4871]: I1007 22:22:48.034044 4871 generic.go:334] "Generic (PLEG): container finished" podID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerID="996bbd2613e94654fa806bbb1646445a1b7f275c05f325fa86053fbc2db267d8" exitCode=0 Oct 07 22:22:48 crc kubenswrapper[4871]: I1007 22:22:48.034272 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xs7cg" event={"ID":"df014e0f-2f9f-4064-b7a8-db411cbe10b8","Type":"ContainerDied","Data":"996bbd2613e94654fa806bbb1646445a1b7f275c05f325fa86053fbc2db267d8"} Oct 07 22:22:49 crc kubenswrapper[4871]: I1007 22:22:49.962145 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-tw5sf" podUID="cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" containerName="console" containerID="cri-o://7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a" gracePeriod=15 Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.052258 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xs7cg" event={"ID":"df014e0f-2f9f-4064-b7a8-db411cbe10b8","Type":"ContainerStarted","Data":"5a02f77f21cb39948f94f0b8b854513d3107d441e01009b3c6bc5e46d3369d1e"} Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.393402 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-tw5sf_cc5b34f7-63fd-4d8d-ba64-5d2954578cb0/console/0.log" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.393752 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.414155 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xs7cg" podStartSLOduration=3.325590386 podStartE2EDuration="6.414129642s" podCreationTimestamp="2025-10-07 22:22:44 +0000 UTC" firstStartedPulling="2025-10-07 22:22:46.018174248 +0000 UTC m=+839.820872361" lastFinishedPulling="2025-10-07 22:22:49.106713524 +0000 UTC m=+842.909411617" observedRunningTime="2025-10-07 22:22:50.07300368 +0000 UTC m=+843.875701753" watchObservedRunningTime="2025-10-07 22:22:50.414129642 +0000 UTC m=+844.216827715" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.563255 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-config\") pod \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.563329 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-service-ca\") pod \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.563439 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvhl7\" (UniqueName: \"kubernetes.io/projected/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-kube-api-access-nvhl7\") pod \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.564277 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-config" (OuterVolumeSpecName: "console-config") pod "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" (UID: "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.564499 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-service-ca" (OuterVolumeSpecName: "service-ca") pod "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" (UID: "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.564819 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-oauth-config\") pod \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.564916 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-trusted-ca-bundle\") pod \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.564961 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-serving-cert\") pod \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.564986 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-oauth-serving-cert\") pod \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\" (UID: \"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0\") " Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.565298 4871 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.565323 4871 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.565561 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" (UID: "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.565577 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" (UID: "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.571047 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-kube-api-access-nvhl7" (OuterVolumeSpecName: "kube-api-access-nvhl7") pod "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" (UID: "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0"). InnerVolumeSpecName "kube-api-access-nvhl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.572061 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" (UID: "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.572306 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" (UID: "cc5b34f7-63fd-4d8d-ba64-5d2954578cb0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.666183 4871 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.666222 4871 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.666235 4871 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.666247 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvhl7\" (UniqueName: \"kubernetes.io/projected/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-kube-api-access-nvhl7\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:50 crc kubenswrapper[4871]: I1007 22:22:50.666261 4871 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.063295 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-tw5sf_cc5b34f7-63fd-4d8d-ba64-5d2954578cb0/console/0.log" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.063361 4871 generic.go:334] "Generic (PLEG): container finished" podID="cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" containerID="7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a" exitCode=2 Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.063475 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tw5sf" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.063442 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tw5sf" event={"ID":"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0","Type":"ContainerDied","Data":"7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a"} Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.063599 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tw5sf" event={"ID":"cc5b34f7-63fd-4d8d-ba64-5d2954578cb0","Type":"ContainerDied","Data":"eb24c72fa1fa40f5278749216bd03264ac337c7a54c25f04b190ee69ff201550"} Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.063662 4871 scope.go:117] "RemoveContainer" containerID="7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.107929 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-tw5sf"] Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.110930 4871 scope.go:117] "RemoveContainer" containerID="7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a" Oct 07 22:22:51 crc kubenswrapper[4871]: E1007 22:22:51.111738 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a\": container with ID starting with 7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a not found: ID does not exist" containerID="7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.111858 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a"} err="failed to get container status \"7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a\": rpc error: code = NotFound desc = could not find container \"7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a\": container with ID starting with 7ecc0487a548ed83092bfa73f75b1f1e57bf990cefd96d77ea46e3ea4cea9d4a not found: ID does not exist" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.113644 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-tw5sf"] Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.250740 4871 patch_prober.go:28] interesting pod/console-f9d7485db-tw5sf container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.250851 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-tw5sf" podUID="cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.668063 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww"] Oct 07 22:22:51 crc kubenswrapper[4871]: E1007 22:22:51.668340 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" containerName="console" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.668355 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" containerName="console" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.668468 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" containerName="console" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.669378 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.672445 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.682774 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6tlf\" (UniqueName: \"kubernetes.io/projected/f0bdd976-9611-4335-93b3-6dec20c06076-kube-api-access-v6tlf\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.683290 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.683451 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.683620 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww"] Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.784612 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.784966 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6tlf\" (UniqueName: \"kubernetes.io/projected/f0bdd976-9611-4335-93b3-6dec20c06076-kube-api-access-v6tlf\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.785074 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.785534 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.785646 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.803259 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6tlf\" (UniqueName: \"kubernetes.io/projected/f0bdd976-9611-4335-93b3-6dec20c06076-kube-api-access-v6tlf\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:51 crc kubenswrapper[4871]: I1007 22:22:51.987664 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:52 crc kubenswrapper[4871]: I1007 22:22:52.466738 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww"] Oct 07 22:22:52 crc kubenswrapper[4871]: W1007 22:22:52.483965 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0bdd976_9611_4335_93b3_6dec20c06076.slice/crio-13cac3f14f7019d93964ba9d4cf1eb1074e04b48dadb29622f289badb8fc8c30 WatchSource:0}: Error finding container 13cac3f14f7019d93964ba9d4cf1eb1074e04b48dadb29622f289badb8fc8c30: Status 404 returned error can't find the container with id 13cac3f14f7019d93964ba9d4cf1eb1074e04b48dadb29622f289badb8fc8c30 Oct 07 22:22:52 crc kubenswrapper[4871]: I1007 22:22:52.998189 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc5b34f7-63fd-4d8d-ba64-5d2954578cb0" path="/var/lib/kubelet/pods/cc5b34f7-63fd-4d8d-ba64-5d2954578cb0/volumes" Oct 07 22:22:53 crc kubenswrapper[4871]: I1007 22:22:53.079930 4871 generic.go:334] "Generic (PLEG): container finished" podID="f0bdd976-9611-4335-93b3-6dec20c06076" containerID="0e21bcc41a3ac7dbcf6601fb0f410abbe2c3097a2d4d6fa4901b03f2b6bafada" exitCode=0 Oct 07 22:22:53 crc kubenswrapper[4871]: I1007 22:22:53.080008 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" event={"ID":"f0bdd976-9611-4335-93b3-6dec20c06076","Type":"ContainerDied","Data":"0e21bcc41a3ac7dbcf6601fb0f410abbe2c3097a2d4d6fa4901b03f2b6bafada"} Oct 07 22:22:53 crc kubenswrapper[4871]: I1007 22:22:53.080091 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" event={"ID":"f0bdd976-9611-4335-93b3-6dec20c06076","Type":"ContainerStarted","Data":"13cac3f14f7019d93964ba9d4cf1eb1074e04b48dadb29622f289badb8fc8c30"} Oct 07 22:22:54 crc kubenswrapper[4871]: I1007 22:22:54.867285 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:54 crc kubenswrapper[4871]: I1007 22:22:54.867369 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:54 crc kubenswrapper[4871]: I1007 22:22:54.943223 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:55 crc kubenswrapper[4871]: I1007 22:22:55.167725 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:56 crc kubenswrapper[4871]: I1007 22:22:56.107832 4871 generic.go:334] "Generic (PLEG): container finished" podID="f0bdd976-9611-4335-93b3-6dec20c06076" containerID="9de435f4e3bf5445d983d869247682dc0c99c8d312f75174bc7f6c057fb6049d" exitCode=0 Oct 07 22:22:56 crc kubenswrapper[4871]: I1007 22:22:56.107934 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" event={"ID":"f0bdd976-9611-4335-93b3-6dec20c06076","Type":"ContainerDied","Data":"9de435f4e3bf5445d983d869247682dc0c99c8d312f75174bc7f6c057fb6049d"} Oct 07 22:22:57 crc kubenswrapper[4871]: I1007 22:22:57.119961 4871 generic.go:334] "Generic (PLEG): container finished" podID="f0bdd976-9611-4335-93b3-6dec20c06076" containerID="b39cd19351dffaf118727aa9eedc999d993683ffd8510c3123f16aabc7d5e788" exitCode=0 Oct 07 22:22:57 crc kubenswrapper[4871]: I1007 22:22:57.120023 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" event={"ID":"f0bdd976-9611-4335-93b3-6dec20c06076","Type":"ContainerDied","Data":"b39cd19351dffaf118727aa9eedc999d993683ffd8510c3123f16aabc7d5e788"} Oct 07 22:22:57 crc kubenswrapper[4871]: I1007 22:22:57.690016 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xs7cg"] Oct 07 22:22:57 crc kubenswrapper[4871]: I1007 22:22:57.690405 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xs7cg" podUID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerName="registry-server" containerID="cri-o://5a02f77f21cb39948f94f0b8b854513d3107d441e01009b3c6bc5e46d3369d1e" gracePeriod=2 Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.129933 4871 generic.go:334] "Generic (PLEG): container finished" podID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerID="5a02f77f21cb39948f94f0b8b854513d3107d441e01009b3c6bc5e46d3369d1e" exitCode=0 Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.130091 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xs7cg" event={"ID":"df014e0f-2f9f-4064-b7a8-db411cbe10b8","Type":"ContainerDied","Data":"5a02f77f21cb39948f94f0b8b854513d3107d441e01009b3c6bc5e46d3369d1e"} Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.182071 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.199730 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgqcc\" (UniqueName: \"kubernetes.io/projected/df014e0f-2f9f-4064-b7a8-db411cbe10b8-kube-api-access-lgqcc\") pod \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.200301 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-catalog-content\") pod \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.200406 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-utilities\") pod \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\" (UID: \"df014e0f-2f9f-4064-b7a8-db411cbe10b8\") " Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.202662 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-utilities" (OuterVolumeSpecName: "utilities") pod "df014e0f-2f9f-4064-b7a8-db411cbe10b8" (UID: "df014e0f-2f9f-4064-b7a8-db411cbe10b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.233848 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df014e0f-2f9f-4064-b7a8-db411cbe10b8-kube-api-access-lgqcc" (OuterVolumeSpecName: "kube-api-access-lgqcc") pod "df014e0f-2f9f-4064-b7a8-db411cbe10b8" (UID: "df014e0f-2f9f-4064-b7a8-db411cbe10b8"). InnerVolumeSpecName "kube-api-access-lgqcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.302630 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgqcc\" (UniqueName: \"kubernetes.io/projected/df014e0f-2f9f-4064-b7a8-db411cbe10b8-kube-api-access-lgqcc\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.302667 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.486194 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.504810 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6tlf\" (UniqueName: \"kubernetes.io/projected/f0bdd976-9611-4335-93b3-6dec20c06076-kube-api-access-v6tlf\") pod \"f0bdd976-9611-4335-93b3-6dec20c06076\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.504946 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-bundle\") pod \"f0bdd976-9611-4335-93b3-6dec20c06076\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.505074 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-util\") pod \"f0bdd976-9611-4335-93b3-6dec20c06076\" (UID: \"f0bdd976-9611-4335-93b3-6dec20c06076\") " Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.507920 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-bundle" (OuterVolumeSpecName: "bundle") pod "f0bdd976-9611-4335-93b3-6dec20c06076" (UID: "f0bdd976-9611-4335-93b3-6dec20c06076"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.515931 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0bdd976-9611-4335-93b3-6dec20c06076-kube-api-access-v6tlf" (OuterVolumeSpecName: "kube-api-access-v6tlf") pod "f0bdd976-9611-4335-93b3-6dec20c06076" (UID: "f0bdd976-9611-4335-93b3-6dec20c06076"). InnerVolumeSpecName "kube-api-access-v6tlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.607147 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6tlf\" (UniqueName: \"kubernetes.io/projected/f0bdd976-9611-4335-93b3-6dec20c06076-kube-api-access-v6tlf\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.607199 4871 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.714978 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-util" (OuterVolumeSpecName: "util") pod "f0bdd976-9611-4335-93b3-6dec20c06076" (UID: "f0bdd976-9611-4335-93b3-6dec20c06076"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.728381 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df014e0f-2f9f-4064-b7a8-db411cbe10b8" (UID: "df014e0f-2f9f-4064-b7a8-db411cbe10b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.811496 4871 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f0bdd976-9611-4335-93b3-6dec20c06076-util\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:58 crc kubenswrapper[4871]: I1007 22:22:58.811593 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df014e0f-2f9f-4064-b7a8-db411cbe10b8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.138854 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" event={"ID":"f0bdd976-9611-4335-93b3-6dec20c06076","Type":"ContainerDied","Data":"13cac3f14f7019d93964ba9d4cf1eb1074e04b48dadb29622f289badb8fc8c30"} Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.138924 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww" Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.138937 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13cac3f14f7019d93964ba9d4cf1eb1074e04b48dadb29622f289badb8fc8c30" Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.141168 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xs7cg" event={"ID":"df014e0f-2f9f-4064-b7a8-db411cbe10b8","Type":"ContainerDied","Data":"c76ab46b55d670d67ce7546132571864e8b549db2b07812801e6902e0219cd44"} Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.141207 4871 scope.go:117] "RemoveContainer" containerID="5a02f77f21cb39948f94f0b8b854513d3107d441e01009b3c6bc5e46d3369d1e" Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.141558 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xs7cg" Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.173637 4871 scope.go:117] "RemoveContainer" containerID="996bbd2613e94654fa806bbb1646445a1b7f275c05f325fa86053fbc2db267d8" Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.187741 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xs7cg"] Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.191027 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xs7cg"] Oct 07 22:22:59 crc kubenswrapper[4871]: I1007 22:22:59.199573 4871 scope.go:117] "RemoveContainer" containerID="a46685f20fb40e1e880918c97820e9c2e3e3fcaf6724bc3652b77f5b62226f21" Oct 07 22:23:00 crc kubenswrapper[4871]: I1007 22:23:00.991499 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" path="/var/lib/kubelet/pods/df014e0f-2f9f-4064-b7a8-db411cbe10b8/volumes" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.881626 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb"] Oct 07 22:23:07 crc kubenswrapper[4871]: E1007 22:23:07.882473 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerName="extract-content" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.882487 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerName="extract-content" Oct 07 22:23:07 crc kubenswrapper[4871]: E1007 22:23:07.882496 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerName="registry-server" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.882502 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerName="registry-server" Oct 07 22:23:07 crc kubenswrapper[4871]: E1007 22:23:07.882509 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0bdd976-9611-4335-93b3-6dec20c06076" containerName="pull" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.882515 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0bdd976-9611-4335-93b3-6dec20c06076" containerName="pull" Oct 07 22:23:07 crc kubenswrapper[4871]: E1007 22:23:07.882530 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerName="extract-utilities" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.882536 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerName="extract-utilities" Oct 07 22:23:07 crc kubenswrapper[4871]: E1007 22:23:07.882543 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0bdd976-9611-4335-93b3-6dec20c06076" containerName="extract" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.882549 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0bdd976-9611-4335-93b3-6dec20c06076" containerName="extract" Oct 07 22:23:07 crc kubenswrapper[4871]: E1007 22:23:07.882560 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0bdd976-9611-4335-93b3-6dec20c06076" containerName="util" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.882565 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0bdd976-9611-4335-93b3-6dec20c06076" containerName="util" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.882664 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0bdd976-9611-4335-93b3-6dec20c06076" containerName="extract" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.882678 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="df014e0f-2f9f-4064-b7a8-db411cbe10b8" containerName="registry-server" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.883103 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.898832 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.900045 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-52476" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.900269 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.903859 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.908861 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.923069 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb"] Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.964600 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/840548f4-0e99-418d-98d4-400e684775e1-apiservice-cert\") pod \"metallb-operator-controller-manager-b78c56f4c-qs2vb\" (UID: \"840548f4-0e99-418d-98d4-400e684775e1\") " pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.964834 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8h5g\" (UniqueName: \"kubernetes.io/projected/840548f4-0e99-418d-98d4-400e684775e1-kube-api-access-k8h5g\") pod \"metallb-operator-controller-manager-b78c56f4c-qs2vb\" (UID: \"840548f4-0e99-418d-98d4-400e684775e1\") " pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:07 crc kubenswrapper[4871]: I1007 22:23:07.965111 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/840548f4-0e99-418d-98d4-400e684775e1-webhook-cert\") pod \"metallb-operator-controller-manager-b78c56f4c-qs2vb\" (UID: \"840548f4-0e99-418d-98d4-400e684775e1\") " pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.067088 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/840548f4-0e99-418d-98d4-400e684775e1-webhook-cert\") pod \"metallb-operator-controller-manager-b78c56f4c-qs2vb\" (UID: \"840548f4-0e99-418d-98d4-400e684775e1\") " pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.067480 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/840548f4-0e99-418d-98d4-400e684775e1-apiservice-cert\") pod \"metallb-operator-controller-manager-b78c56f4c-qs2vb\" (UID: \"840548f4-0e99-418d-98d4-400e684775e1\") " pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.067571 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8h5g\" (UniqueName: \"kubernetes.io/projected/840548f4-0e99-418d-98d4-400e684775e1-kube-api-access-k8h5g\") pod \"metallb-operator-controller-manager-b78c56f4c-qs2vb\" (UID: \"840548f4-0e99-418d-98d4-400e684775e1\") " pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.075738 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/840548f4-0e99-418d-98d4-400e684775e1-webhook-cert\") pod \"metallb-operator-controller-manager-b78c56f4c-qs2vb\" (UID: \"840548f4-0e99-418d-98d4-400e684775e1\") " pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.076209 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/840548f4-0e99-418d-98d4-400e684775e1-apiservice-cert\") pod \"metallb-operator-controller-manager-b78c56f4c-qs2vb\" (UID: \"840548f4-0e99-418d-98d4-400e684775e1\") " pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.094942 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8h5g\" (UniqueName: \"kubernetes.io/projected/840548f4-0e99-418d-98d4-400e684775e1-kube-api-access-k8h5g\") pod \"metallb-operator-controller-manager-b78c56f4c-qs2vb\" (UID: \"840548f4-0e99-418d-98d4-400e684775e1\") " pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.198628 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.228505 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv"] Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.229500 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.235252 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-gw98r" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.235252 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.235259 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.247492 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv"] Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.273168 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdtpg\" (UniqueName: \"kubernetes.io/projected/ad41768e-7589-4b4e-8193-2b48da644e53-kube-api-access-xdtpg\") pod \"metallb-operator-webhook-server-76fdd8f567-7nxxv\" (UID: \"ad41768e-7589-4b4e-8193-2b48da644e53\") " pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.273247 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ad41768e-7589-4b4e-8193-2b48da644e53-webhook-cert\") pod \"metallb-operator-webhook-server-76fdd8f567-7nxxv\" (UID: \"ad41768e-7589-4b4e-8193-2b48da644e53\") " pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.273321 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ad41768e-7589-4b4e-8193-2b48da644e53-apiservice-cert\") pod \"metallb-operator-webhook-server-76fdd8f567-7nxxv\" (UID: \"ad41768e-7589-4b4e-8193-2b48da644e53\") " pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.374699 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdtpg\" (UniqueName: \"kubernetes.io/projected/ad41768e-7589-4b4e-8193-2b48da644e53-kube-api-access-xdtpg\") pod \"metallb-operator-webhook-server-76fdd8f567-7nxxv\" (UID: \"ad41768e-7589-4b4e-8193-2b48da644e53\") " pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.375221 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ad41768e-7589-4b4e-8193-2b48da644e53-webhook-cert\") pod \"metallb-operator-webhook-server-76fdd8f567-7nxxv\" (UID: \"ad41768e-7589-4b4e-8193-2b48da644e53\") " pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.375301 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ad41768e-7589-4b4e-8193-2b48da644e53-apiservice-cert\") pod \"metallb-operator-webhook-server-76fdd8f567-7nxxv\" (UID: \"ad41768e-7589-4b4e-8193-2b48da644e53\") " pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.386007 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ad41768e-7589-4b4e-8193-2b48da644e53-webhook-cert\") pod \"metallb-operator-webhook-server-76fdd8f567-7nxxv\" (UID: \"ad41768e-7589-4b4e-8193-2b48da644e53\") " pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.394258 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdtpg\" (UniqueName: \"kubernetes.io/projected/ad41768e-7589-4b4e-8193-2b48da644e53-kube-api-access-xdtpg\") pod \"metallb-operator-webhook-server-76fdd8f567-7nxxv\" (UID: \"ad41768e-7589-4b4e-8193-2b48da644e53\") " pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.403190 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ad41768e-7589-4b4e-8193-2b48da644e53-apiservice-cert\") pod \"metallb-operator-webhook-server-76fdd8f567-7nxxv\" (UID: \"ad41768e-7589-4b4e-8193-2b48da644e53\") " pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.473579 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb"] Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.598961 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:08 crc kubenswrapper[4871]: I1007 22:23:08.823188 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv"] Oct 07 22:23:08 crc kubenswrapper[4871]: W1007 22:23:08.834001 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad41768e_7589_4b4e_8193_2b48da644e53.slice/crio-840009e877f44168f9e457ff746910e7f773243d3f8a3fc31edd56a862da998e WatchSource:0}: Error finding container 840009e877f44168f9e457ff746910e7f773243d3f8a3fc31edd56a862da998e: Status 404 returned error can't find the container with id 840009e877f44168f9e457ff746910e7f773243d3f8a3fc31edd56a862da998e Oct 07 22:23:09 crc kubenswrapper[4871]: I1007 22:23:09.203840 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" event={"ID":"840548f4-0e99-418d-98d4-400e684775e1","Type":"ContainerStarted","Data":"6d032358f313df0f87a101dc1686d9d757429bff51d96ebb776e628d7ace7e53"} Oct 07 22:23:09 crc kubenswrapper[4871]: I1007 22:23:09.205642 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" event={"ID":"ad41768e-7589-4b4e-8193-2b48da644e53","Type":"ContainerStarted","Data":"840009e877f44168f9e457ff746910e7f773243d3f8a3fc31edd56a862da998e"} Oct 07 22:23:14 crc kubenswrapper[4871]: I1007 22:23:14.254244 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" event={"ID":"840548f4-0e99-418d-98d4-400e684775e1","Type":"ContainerStarted","Data":"f211701b17990c7a7fd8e195525dc8b06f48c16c852a15c2f228556168752483"} Oct 07 22:23:14 crc kubenswrapper[4871]: I1007 22:23:14.255216 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:14 crc kubenswrapper[4871]: I1007 22:23:14.293252 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" podStartSLOduration=2.723669915 podStartE2EDuration="7.293228475s" podCreationTimestamp="2025-10-07 22:23:07 +0000 UTC" firstStartedPulling="2025-10-07 22:23:08.489893465 +0000 UTC m=+862.292591538" lastFinishedPulling="2025-10-07 22:23:13.059452035 +0000 UTC m=+866.862150098" observedRunningTime="2025-10-07 22:23:14.282507713 +0000 UTC m=+868.085205786" watchObservedRunningTime="2025-10-07 22:23:14.293228475 +0000 UTC m=+868.095926548" Oct 07 22:23:16 crc kubenswrapper[4871]: I1007 22:23:16.273468 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" event={"ID":"ad41768e-7589-4b4e-8193-2b48da644e53","Type":"ContainerStarted","Data":"b42600486df8e5a2172e5237158218a70036948a5e432f507698a5f4cb78cea1"} Oct 07 22:23:16 crc kubenswrapper[4871]: I1007 22:23:16.319149 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" podStartSLOduration=1.826823039 podStartE2EDuration="8.319121301s" podCreationTimestamp="2025-10-07 22:23:08 +0000 UTC" firstStartedPulling="2025-10-07 22:23:08.838121924 +0000 UTC m=+862.640819997" lastFinishedPulling="2025-10-07 22:23:15.330420186 +0000 UTC m=+869.133118259" observedRunningTime="2025-10-07 22:23:16.315967288 +0000 UTC m=+870.118665401" watchObservedRunningTime="2025-10-07 22:23:16.319121301 +0000 UTC m=+870.121819374" Oct 07 22:23:17 crc kubenswrapper[4871]: I1007 22:23:17.283431 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:28 crc kubenswrapper[4871]: I1007 22:23:28.608103 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-76fdd8f567-7nxxv" Oct 07 22:23:35 crc kubenswrapper[4871]: I1007 22:23:35.512925 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:23:35 crc kubenswrapper[4871]: I1007 22:23:35.513876 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.202425 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-b78c56f4c-qs2vb" Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.918546 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m"] Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.920470 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.922769 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.924189 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-ckhc9" Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.932853 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m"] Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.937456 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ksdzm"] Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.940270 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.942159 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 07 22:23:48 crc kubenswrapper[4871]: I1007 22:23:48.942966 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.014535 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tltx\" (UniqueName: \"kubernetes.io/projected/f2649d22-c9ed-4d54-b85f-7739a2b50db2-kube-api-access-5tltx\") pod \"frr-k8s-webhook-server-64bf5d555-2fc4m\" (UID: \"f2649d22-c9ed-4d54-b85f-7739a2b50db2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.014919 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f2649d22-c9ed-4d54-b85f-7739a2b50db2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-2fc4m\" (UID: \"f2649d22-c9ed-4d54-b85f-7739a2b50db2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.026628 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-4hjwr"] Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.027515 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.031364 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.031566 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.031666 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.031903 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-7922n" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.049637 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-gm89v"] Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.050757 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.053215 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.072491 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-gm89v"] Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.116917 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tltx\" (UniqueName: \"kubernetes.io/projected/f2649d22-c9ed-4d54-b85f-7739a2b50db2-kube-api-access-5tltx\") pod \"frr-k8s-webhook-server-64bf5d555-2fc4m\" (UID: \"f2649d22-c9ed-4d54-b85f-7739a2b50db2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117021 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f2649d22-c9ed-4d54-b85f-7739a2b50db2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-2fc4m\" (UID: \"f2649d22-c9ed-4d54-b85f-7739a2b50db2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117092 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-frr-startup\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117128 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-reloader\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117211 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-frr-sockets\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117289 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mk4p\" (UniqueName: \"kubernetes.io/projected/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-kube-api-access-4mk4p\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117342 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-metrics\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117364 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-metrics-certs\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117404 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-memberlist\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117432 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-frr-conf\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117465 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-metallb-excludel2\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117490 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4zbr\" (UniqueName: \"kubernetes.io/projected/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-kube-api-access-b4zbr\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.117543 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-metrics-certs\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.126023 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f2649d22-c9ed-4d54-b85f-7739a2b50db2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-2fc4m\" (UID: \"f2649d22-c9ed-4d54-b85f-7739a2b50db2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.138526 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tltx\" (UniqueName: \"kubernetes.io/projected/f2649d22-c9ed-4d54-b85f-7739a2b50db2-kube-api-access-5tltx\") pod \"frr-k8s-webhook-server-64bf5d555-2fc4m\" (UID: \"f2649d22-c9ed-4d54-b85f-7739a2b50db2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219157 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-frr-sockets\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219241 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mk4p\" (UniqueName: \"kubernetes.io/projected/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-kube-api-access-4mk4p\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219279 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578-cert\") pod \"controller-68d546b9d8-gm89v\" (UID: \"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578\") " pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219310 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swb4r\" (UniqueName: \"kubernetes.io/projected/00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578-kube-api-access-swb4r\") pod \"controller-68d546b9d8-gm89v\" (UID: \"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578\") " pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219336 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-metrics\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219357 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-metrics-certs\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219382 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-memberlist\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219402 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-frr-conf\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219425 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578-metrics-certs\") pod \"controller-68d546b9d8-gm89v\" (UID: \"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578\") " pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219455 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-metallb-excludel2\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219477 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4zbr\" (UniqueName: \"kubernetes.io/projected/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-kube-api-access-b4zbr\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219511 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-metrics-certs\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219554 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-frr-startup\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.219576 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-reloader\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.220081 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-metrics\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.220135 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-reloader\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.220383 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-frr-sockets\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.224025 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-metallb-excludel2\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.224510 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-frr-conf\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: E1007 22:23:49.224876 4871 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 07 22:23:49 crc kubenswrapper[4871]: E1007 22:23:49.224945 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-memberlist podName:f494e98b-a5ac-4f64-8f30-4da19c9bc9f7 nodeName:}" failed. No retries permitted until 2025-10-07 22:23:49.724925988 +0000 UTC m=+903.527624061 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-memberlist") pod "speaker-4hjwr" (UID: "f494e98b-a5ac-4f64-8f30-4da19c9bc9f7") : secret "metallb-memberlist" not found Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.225957 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-frr-startup\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.227867 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-metrics-certs\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.232415 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-metrics-certs\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.236053 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.239268 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mk4p\" (UniqueName: \"kubernetes.io/projected/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-kube-api-access-4mk4p\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.243875 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4zbr\" (UniqueName: \"kubernetes.io/projected/7c3493d5-0f90-4c6b-8ebc-ab2a78082037-kube-api-access-b4zbr\") pod \"frr-k8s-ksdzm\" (UID: \"7c3493d5-0f90-4c6b-8ebc-ab2a78082037\") " pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.258759 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.321692 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578-cert\") pod \"controller-68d546b9d8-gm89v\" (UID: \"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578\") " pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.322204 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swb4r\" (UniqueName: \"kubernetes.io/projected/00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578-kube-api-access-swb4r\") pod \"controller-68d546b9d8-gm89v\" (UID: \"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578\") " pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.322266 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578-metrics-certs\") pod \"controller-68d546b9d8-gm89v\" (UID: \"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578\") " pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.324962 4871 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.327619 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578-metrics-certs\") pod \"controller-68d546b9d8-gm89v\" (UID: \"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578\") " pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.342469 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578-cert\") pod \"controller-68d546b9d8-gm89v\" (UID: \"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578\") " pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.345772 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swb4r\" (UniqueName: \"kubernetes.io/projected/00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578-kube-api-access-swb4r\") pod \"controller-68d546b9d8-gm89v\" (UID: \"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578\") " pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.370561 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.461710 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m"] Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.528882 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerStarted","Data":"8f5777d319c5f7fe307058ea1ace431b2410850c2cdb582731a31990555ce88d"} Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.530251 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" event={"ID":"f2649d22-c9ed-4d54-b85f-7739a2b50db2","Type":"ContainerStarted","Data":"365fc49b309a32eaa0e65d091e7a1ad19430c203e036f5a2adcc0b07171bb905"} Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.625231 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-gm89v"] Oct 07 22:23:49 crc kubenswrapper[4871]: I1007 22:23:49.727176 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-memberlist\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:49 crc kubenswrapper[4871]: E1007 22:23:49.727357 4871 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 07 22:23:49 crc kubenswrapper[4871]: E1007 22:23:49.727460 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-memberlist podName:f494e98b-a5ac-4f64-8f30-4da19c9bc9f7 nodeName:}" failed. No retries permitted until 2025-10-07 22:23:50.727435755 +0000 UTC m=+904.530133828 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-memberlist") pod "speaker-4hjwr" (UID: "f494e98b-a5ac-4f64-8f30-4da19c9bc9f7") : secret "metallb-memberlist" not found Oct 07 22:23:50 crc kubenswrapper[4871]: I1007 22:23:50.539402 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-gm89v" event={"ID":"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578","Type":"ContainerStarted","Data":"c253cd843d494b1d00c927de0dae2a1b00fd4a86adb69a7b367f0b450d0d4211"} Oct 07 22:23:50 crc kubenswrapper[4871]: I1007 22:23:50.539968 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-gm89v" event={"ID":"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578","Type":"ContainerStarted","Data":"44e72666bad2aabda291c58e86b0ff6cc65de962383352ad4c48aab1a167142f"} Oct 07 22:23:50 crc kubenswrapper[4871]: I1007 22:23:50.539983 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-gm89v" event={"ID":"00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578","Type":"ContainerStarted","Data":"8cf6bd8d7a346fb1cc2d38231c3d29d68a6243407baad4f0428bbcd99f3f9ca0"} Oct 07 22:23:50 crc kubenswrapper[4871]: I1007 22:23:50.539998 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:50 crc kubenswrapper[4871]: I1007 22:23:50.562349 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-gm89v" podStartSLOduration=1.562322555 podStartE2EDuration="1.562322555s" podCreationTimestamp="2025-10-07 22:23:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:23:50.559496461 +0000 UTC m=+904.362194574" watchObservedRunningTime="2025-10-07 22:23:50.562322555 +0000 UTC m=+904.365020638" Oct 07 22:23:50 crc kubenswrapper[4871]: I1007 22:23:50.741432 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-memberlist\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:50 crc kubenswrapper[4871]: I1007 22:23:50.749905 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f494e98b-a5ac-4f64-8f30-4da19c9bc9f7-memberlist\") pod \"speaker-4hjwr\" (UID: \"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7\") " pod="metallb-system/speaker-4hjwr" Oct 07 22:23:50 crc kubenswrapper[4871]: I1007 22:23:50.843126 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4hjwr" Oct 07 22:23:50 crc kubenswrapper[4871]: W1007 22:23:50.871323 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf494e98b_a5ac_4f64_8f30_4da19c9bc9f7.slice/crio-5ea9f6dffea97f056ebb1549820ac6d1fb71da67578451e23dce055020f58586 WatchSource:0}: Error finding container 5ea9f6dffea97f056ebb1549820ac6d1fb71da67578451e23dce055020f58586: Status 404 returned error can't find the container with id 5ea9f6dffea97f056ebb1549820ac6d1fb71da67578451e23dce055020f58586 Oct 07 22:23:51 crc kubenswrapper[4871]: I1007 22:23:51.549997 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4hjwr" event={"ID":"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7","Type":"ContainerStarted","Data":"d241bd5b65bb9946cf29677708783ed154afd260379b2b3f1252eb341584c60f"} Oct 07 22:23:51 crc kubenswrapper[4871]: I1007 22:23:51.550544 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4hjwr" event={"ID":"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7","Type":"ContainerStarted","Data":"7f9fa8fb3c8db92b05c17bc33d3867cd7ae64d2b05b4145b976bb29abf976a9c"} Oct 07 22:23:51 crc kubenswrapper[4871]: I1007 22:23:51.550570 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4hjwr" event={"ID":"f494e98b-a5ac-4f64-8f30-4da19c9bc9f7","Type":"ContainerStarted","Data":"5ea9f6dffea97f056ebb1549820ac6d1fb71da67578451e23dce055020f58586"} Oct 07 22:23:51 crc kubenswrapper[4871]: I1007 22:23:51.550917 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-4hjwr" Oct 07 22:23:51 crc kubenswrapper[4871]: I1007 22:23:51.579548 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-4hjwr" podStartSLOduration=2.57951195 podStartE2EDuration="2.57951195s" podCreationTimestamp="2025-10-07 22:23:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:23:51.578458322 +0000 UTC m=+905.381156435" watchObservedRunningTime="2025-10-07 22:23:51.57951195 +0000 UTC m=+905.382210063" Oct 07 22:23:59 crc kubenswrapper[4871]: I1007 22:23:59.376680 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-gm89v" Oct 07 22:23:59 crc kubenswrapper[4871]: I1007 22:23:59.620363 4871 generic.go:334] "Generic (PLEG): container finished" podID="7c3493d5-0f90-4c6b-8ebc-ab2a78082037" containerID="90a8045c0818b5cff6625035995c9602303ef49c6c02105fcfa269891748ceb4" exitCode=0 Oct 07 22:23:59 crc kubenswrapper[4871]: I1007 22:23:59.620485 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerDied","Data":"90a8045c0818b5cff6625035995c9602303ef49c6c02105fcfa269891748ceb4"} Oct 07 22:23:59 crc kubenswrapper[4871]: I1007 22:23:59.623241 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" event={"ID":"f2649d22-c9ed-4d54-b85f-7739a2b50db2","Type":"ContainerStarted","Data":"0282505b8a53f68eccf0755996b9566ef72607f78247a6254cb649ec9a748916"} Oct 07 22:23:59 crc kubenswrapper[4871]: I1007 22:23:59.623407 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:23:59 crc kubenswrapper[4871]: I1007 22:23:59.719300 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" podStartSLOduration=2.516739526 podStartE2EDuration="11.719278544s" podCreationTimestamp="2025-10-07 22:23:48 +0000 UTC" firstStartedPulling="2025-10-07 22:23:49.473825015 +0000 UTC m=+903.276523078" lastFinishedPulling="2025-10-07 22:23:58.676364023 +0000 UTC m=+912.479062096" observedRunningTime="2025-10-07 22:23:59.715385732 +0000 UTC m=+913.518083805" watchObservedRunningTime="2025-10-07 22:23:59.719278544 +0000 UTC m=+913.521976617" Oct 07 22:24:00 crc kubenswrapper[4871]: I1007 22:24:00.635877 4871 generic.go:334] "Generic (PLEG): container finished" podID="7c3493d5-0f90-4c6b-8ebc-ab2a78082037" containerID="93a467b8f03a1c813eecad4e249d5303fe391a5c23b4b6a9af10623e40f0ddaf" exitCode=0 Oct 07 22:24:00 crc kubenswrapper[4871]: I1007 22:24:00.635999 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerDied","Data":"93a467b8f03a1c813eecad4e249d5303fe391a5c23b4b6a9af10623e40f0ddaf"} Oct 07 22:24:01 crc kubenswrapper[4871]: I1007 22:24:01.649503 4871 generic.go:334] "Generic (PLEG): container finished" podID="7c3493d5-0f90-4c6b-8ebc-ab2a78082037" containerID="c18b502afa348042325ed397fd49afd9f4f0b897f05466716f65e7d2d8b32260" exitCode=0 Oct 07 22:24:01 crc kubenswrapper[4871]: I1007 22:24:01.649596 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerDied","Data":"c18b502afa348042325ed397fd49afd9f4f0b897f05466716f65e7d2d8b32260"} Oct 07 22:24:02 crc kubenswrapper[4871]: I1007 22:24:02.665799 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerStarted","Data":"de1a4c71acb3e4bd97b2bdd1ce44b795b0a20057b671b83ba662406785ca15c9"} Oct 07 22:24:02 crc kubenswrapper[4871]: I1007 22:24:02.666164 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerStarted","Data":"44a41c61553f806aa055fd7c88f9f27f93949fe8fee157b116cee988e9091917"} Oct 07 22:24:02 crc kubenswrapper[4871]: I1007 22:24:02.666177 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerStarted","Data":"6f89524b03284e81662f52fbf973c2a1b6b432fb5b84bbd2a80af1d681d5b84c"} Oct 07 22:24:02 crc kubenswrapper[4871]: I1007 22:24:02.666186 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerStarted","Data":"ee65c2cb4498b88aaf9be2bc18140526791ce02407f13ee5f0b6909bfae2fd2a"} Oct 07 22:24:02 crc kubenswrapper[4871]: I1007 22:24:02.666196 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerStarted","Data":"d44f31e2fe98e1167de12dca5593e6de227f9d65112a479d2d7ebf92d8fcc389"} Oct 07 22:24:03 crc kubenswrapper[4871]: I1007 22:24:03.678683 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ksdzm" event={"ID":"7c3493d5-0f90-4c6b-8ebc-ab2a78082037","Type":"ContainerStarted","Data":"1b6942c07d0688a6f9225b485a72f424be3a407c161ee6b91a887f78e7626399"} Oct 07 22:24:03 crc kubenswrapper[4871]: I1007 22:24:03.679233 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:24:04 crc kubenswrapper[4871]: I1007 22:24:04.260067 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:24:04 crc kubenswrapper[4871]: I1007 22:24:04.328614 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:24:04 crc kubenswrapper[4871]: I1007 22:24:04.361721 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ksdzm" podStartSLOduration=7.065265472 podStartE2EDuration="16.361662578s" podCreationTimestamp="2025-10-07 22:23:48 +0000 UTC" firstStartedPulling="2025-10-07 22:23:49.414439923 +0000 UTC m=+903.217137996" lastFinishedPulling="2025-10-07 22:23:58.710836989 +0000 UTC m=+912.513535102" observedRunningTime="2025-10-07 22:24:03.72431627 +0000 UTC m=+917.527014383" watchObservedRunningTime="2025-10-07 22:24:04.361662578 +0000 UTC m=+918.164360691" Oct 07 22:24:05 crc kubenswrapper[4871]: I1007 22:24:05.512573 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:24:05 crc kubenswrapper[4871]: I1007 22:24:05.512679 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:24:09 crc kubenswrapper[4871]: I1007 22:24:09.243438 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-2fc4m" Oct 07 22:24:10 crc kubenswrapper[4871]: I1007 22:24:10.847641 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-4hjwr" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.459370 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt"] Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.460710 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.463067 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.485329 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt"] Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.503951 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wknf7\" (UniqueName: \"kubernetes.io/projected/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-kube-api-access-wknf7\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.504025 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.504331 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.606240 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wknf7\" (UniqueName: \"kubernetes.io/projected/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-kube-api-access-wknf7\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.606333 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.606375 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.607003 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.607061 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.628881 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wknf7\" (UniqueName: \"kubernetes.io/projected/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-kube-api-access-wknf7\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:12 crc kubenswrapper[4871]: I1007 22:24:12.778785 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:13 crc kubenswrapper[4871]: I1007 22:24:13.077587 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt"] Oct 07 22:24:13 crc kubenswrapper[4871]: I1007 22:24:13.763544 4871 generic.go:334] "Generic (PLEG): container finished" podID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerID="cb76a4755a01328b3c963f6ba01d62e7a1676c90c2181b5170442f22a712c430" exitCode=0 Oct 07 22:24:13 crc kubenswrapper[4871]: I1007 22:24:13.763619 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" event={"ID":"fe65ed5d-7ec6-47e4-a99a-9c214d055a33","Type":"ContainerDied","Data":"cb76a4755a01328b3c963f6ba01d62e7a1676c90c2181b5170442f22a712c430"} Oct 07 22:24:13 crc kubenswrapper[4871]: I1007 22:24:13.763717 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" event={"ID":"fe65ed5d-7ec6-47e4-a99a-9c214d055a33","Type":"ContainerStarted","Data":"5b820f85cfbe1f5edaf4b87589d485a3f13baad59aed6884a3daa3d97f6a6f56"} Oct 07 22:24:18 crc kubenswrapper[4871]: I1007 22:24:18.814865 4871 generic.go:334] "Generic (PLEG): container finished" podID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerID="b42ac83a42dd46c1f4eba545be091538467a2c2b8be4142c3b987abc2910f1e0" exitCode=0 Oct 07 22:24:18 crc kubenswrapper[4871]: I1007 22:24:18.814993 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" event={"ID":"fe65ed5d-7ec6-47e4-a99a-9c214d055a33","Type":"ContainerDied","Data":"b42ac83a42dd46c1f4eba545be091538467a2c2b8be4142c3b987abc2910f1e0"} Oct 07 22:24:19 crc kubenswrapper[4871]: I1007 22:24:19.262780 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ksdzm" Oct 07 22:24:19 crc kubenswrapper[4871]: I1007 22:24:19.834010 4871 generic.go:334] "Generic (PLEG): container finished" podID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerID="f60706c94c36b0acaeafd1bef93d45476775f1ee28de311612cc570711f263cd" exitCode=0 Oct 07 22:24:19 crc kubenswrapper[4871]: I1007 22:24:19.834085 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" event={"ID":"fe65ed5d-7ec6-47e4-a99a-9c214d055a33","Type":"ContainerDied","Data":"f60706c94c36b0acaeafd1bef93d45476775f1ee28de311612cc570711f263cd"} Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.154627 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.240156 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wknf7\" (UniqueName: \"kubernetes.io/projected/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-kube-api-access-wknf7\") pod \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.240267 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-util\") pod \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.240333 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-bundle\") pod \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\" (UID: \"fe65ed5d-7ec6-47e4-a99a-9c214d055a33\") " Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.241621 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-bundle" (OuterVolumeSpecName: "bundle") pod "fe65ed5d-7ec6-47e4-a99a-9c214d055a33" (UID: "fe65ed5d-7ec6-47e4-a99a-9c214d055a33"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.247370 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-kube-api-access-wknf7" (OuterVolumeSpecName: "kube-api-access-wknf7") pod "fe65ed5d-7ec6-47e4-a99a-9c214d055a33" (UID: "fe65ed5d-7ec6-47e4-a99a-9c214d055a33"). InnerVolumeSpecName "kube-api-access-wknf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.255348 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-util" (OuterVolumeSpecName: "util") pod "fe65ed5d-7ec6-47e4-a99a-9c214d055a33" (UID: "fe65ed5d-7ec6-47e4-a99a-9c214d055a33"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.342552 4871 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-util\") on node \"crc\" DevicePath \"\"" Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.342608 4871 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.342628 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wknf7\" (UniqueName: \"kubernetes.io/projected/fe65ed5d-7ec6-47e4-a99a-9c214d055a33-kube-api-access-wknf7\") on node \"crc\" DevicePath \"\"" Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.852941 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" event={"ID":"fe65ed5d-7ec6-47e4-a99a-9c214d055a33","Type":"ContainerDied","Data":"5b820f85cfbe1f5edaf4b87589d485a3f13baad59aed6884a3daa3d97f6a6f56"} Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.852993 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b820f85cfbe1f5edaf4b87589d485a3f13baad59aed6884a3daa3d97f6a6f56" Oct 07 22:24:21 crc kubenswrapper[4871]: I1007 22:24:21.853047 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.769916 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x"] Oct 07 22:24:32 crc kubenswrapper[4871]: E1007 22:24:32.771072 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerName="extract" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.771090 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerName="extract" Oct 07 22:24:32 crc kubenswrapper[4871]: E1007 22:24:32.771102 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerName="pull" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.771110 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerName="pull" Oct 07 22:24:32 crc kubenswrapper[4871]: E1007 22:24:32.771127 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerName="util" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.771139 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerName="util" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.771295 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe65ed5d-7ec6-47e4-a99a-9c214d055a33" containerName="extract" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.771784 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.774493 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.774884 4871 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-txpg4" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.775434 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.790544 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x"] Oct 07 22:24:32 crc kubenswrapper[4871]: I1007 22:24:32.936422 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdw26\" (UniqueName: \"kubernetes.io/projected/53b000b9-c0a4-4402-9f9a-c3b1a293bfc0-kube-api-access-sdw26\") pod \"cert-manager-operator-controller-manager-57cd46d6d-wpw8x\" (UID: \"53b000b9-c0a4-4402-9f9a-c3b1a293bfc0\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x" Oct 07 22:24:33 crc kubenswrapper[4871]: I1007 22:24:33.037782 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdw26\" (UniqueName: \"kubernetes.io/projected/53b000b9-c0a4-4402-9f9a-c3b1a293bfc0-kube-api-access-sdw26\") pod \"cert-manager-operator-controller-manager-57cd46d6d-wpw8x\" (UID: \"53b000b9-c0a4-4402-9f9a-c3b1a293bfc0\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x" Oct 07 22:24:33 crc kubenswrapper[4871]: I1007 22:24:33.074255 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdw26\" (UniqueName: \"kubernetes.io/projected/53b000b9-c0a4-4402-9f9a-c3b1a293bfc0-kube-api-access-sdw26\") pod \"cert-manager-operator-controller-manager-57cd46d6d-wpw8x\" (UID: \"53b000b9-c0a4-4402-9f9a-c3b1a293bfc0\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x" Oct 07 22:24:33 crc kubenswrapper[4871]: I1007 22:24:33.104408 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x" Oct 07 22:24:33 crc kubenswrapper[4871]: I1007 22:24:33.376392 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x"] Oct 07 22:24:33 crc kubenswrapper[4871]: W1007 22:24:33.383027 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53b000b9_c0a4_4402_9f9a_c3b1a293bfc0.slice/crio-f477c13fc9a5b6b32995a7e2117a55480fb5c7a4a18934e3d18eff60049db670 WatchSource:0}: Error finding container f477c13fc9a5b6b32995a7e2117a55480fb5c7a4a18934e3d18eff60049db670: Status 404 returned error can't find the container with id f477c13fc9a5b6b32995a7e2117a55480fb5c7a4a18934e3d18eff60049db670 Oct 07 22:24:33 crc kubenswrapper[4871]: I1007 22:24:33.950906 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x" event={"ID":"53b000b9-c0a4-4402-9f9a-c3b1a293bfc0","Type":"ContainerStarted","Data":"f477c13fc9a5b6b32995a7e2117a55480fb5c7a4a18934e3d18eff60049db670"} Oct 07 22:24:35 crc kubenswrapper[4871]: I1007 22:24:35.512753 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:24:35 crc kubenswrapper[4871]: I1007 22:24:35.512871 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:24:35 crc kubenswrapper[4871]: I1007 22:24:35.512943 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:24:35 crc kubenswrapper[4871]: I1007 22:24:35.513637 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fbc94adc3f0888dbb1bfd5173d9407ea8c5c758052b4cd3dbe78d9b527973c4"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:24:35 crc kubenswrapper[4871]: I1007 22:24:35.513731 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://5fbc94adc3f0888dbb1bfd5173d9407ea8c5c758052b4cd3dbe78d9b527973c4" gracePeriod=600 Oct 07 22:24:35 crc kubenswrapper[4871]: I1007 22:24:35.968282 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="5fbc94adc3f0888dbb1bfd5173d9407ea8c5c758052b4cd3dbe78d9b527973c4" exitCode=0 Oct 07 22:24:35 crc kubenswrapper[4871]: I1007 22:24:35.968498 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"5fbc94adc3f0888dbb1bfd5173d9407ea8c5c758052b4cd3dbe78d9b527973c4"} Oct 07 22:24:35 crc kubenswrapper[4871]: I1007 22:24:35.968756 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"e0ffdee3d72686448819835e9ee3e3402bc2e2e3b339676f30e5a058e9fac0be"} Oct 07 22:24:35 crc kubenswrapper[4871]: I1007 22:24:35.968817 4871 scope.go:117] "RemoveContainer" containerID="f191ac46661d036cac4f07f7277192def3c4b3a3a733b08403472441f085e8e0" Oct 07 22:24:42 crc kubenswrapper[4871]: I1007 22:24:42.039161 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x" event={"ID":"53b000b9-c0a4-4402-9f9a-c3b1a293bfc0","Type":"ContainerStarted","Data":"7a516312b71e2043083bfa4ab48615c3d7bd40bd784690f02093ae6fcbd2d889"} Oct 07 22:24:42 crc kubenswrapper[4871]: I1007 22:24:42.080333 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-wpw8x" podStartSLOduration=2.033505624 podStartE2EDuration="10.0803039s" podCreationTimestamp="2025-10-07 22:24:32 +0000 UTC" firstStartedPulling="2025-10-07 22:24:33.38720161 +0000 UTC m=+947.189899683" lastFinishedPulling="2025-10-07 22:24:41.433999846 +0000 UTC m=+955.236697959" observedRunningTime="2025-10-07 22:24:42.077327231 +0000 UTC m=+955.880025354" watchObservedRunningTime="2025-10-07 22:24:42.0803039 +0000 UTC m=+955.883002003" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.155606 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-zls94"] Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.157035 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.161632 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.161707 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ftc9\" (UniqueName: \"kubernetes.io/projected/90172f12-e077-4e3a-93a0-115a93588eac-kube-api-access-9ftc9\") pod \"cert-manager-webhook-d969966f-zls94\" (UID: \"90172f12-e077-4e3a-93a0-115a93588eac\") " pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.161807 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90172f12-e077-4e3a-93a0-115a93588eac-bound-sa-token\") pod \"cert-manager-webhook-d969966f-zls94\" (UID: \"90172f12-e077-4e3a-93a0-115a93588eac\") " pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.161835 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.161967 4871 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-227pc" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.171207 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-zls94"] Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.263872 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ftc9\" (UniqueName: \"kubernetes.io/projected/90172f12-e077-4e3a-93a0-115a93588eac-kube-api-access-9ftc9\") pod \"cert-manager-webhook-d969966f-zls94\" (UID: \"90172f12-e077-4e3a-93a0-115a93588eac\") " pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.264037 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90172f12-e077-4e3a-93a0-115a93588eac-bound-sa-token\") pod \"cert-manager-webhook-d969966f-zls94\" (UID: \"90172f12-e077-4e3a-93a0-115a93588eac\") " pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.291514 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90172f12-e077-4e3a-93a0-115a93588eac-bound-sa-token\") pod \"cert-manager-webhook-d969966f-zls94\" (UID: \"90172f12-e077-4e3a-93a0-115a93588eac\") " pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.294034 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ftc9\" (UniqueName: \"kubernetes.io/projected/90172f12-e077-4e3a-93a0-115a93588eac-kube-api-access-9ftc9\") pod \"cert-manager-webhook-d969966f-zls94\" (UID: \"90172f12-e077-4e3a-93a0-115a93588eac\") " pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.499289 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:24:45 crc kubenswrapper[4871]: I1007 22:24:45.998694 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-zls94"] Oct 07 22:24:46 crc kubenswrapper[4871]: I1007 22:24:46.063731 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-zls94" event={"ID":"90172f12-e077-4e3a-93a0-115a93588eac","Type":"ContainerStarted","Data":"3da15c361d6d950e5be37363eb31d4dac1d4f70cd8598e6e077289e81039a675"} Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.588567 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp"] Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.589939 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.598123 4871 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-bjb62" Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.601325 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp"] Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.617850 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7897ec6d-8c3f-4f9b-93b2-489e21dc01af-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-9x9xp\" (UID: \"7897ec6d-8c3f-4f9b-93b2-489e21dc01af\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.617903 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smpgz\" (UniqueName: \"kubernetes.io/projected/7897ec6d-8c3f-4f9b-93b2-489e21dc01af-kube-api-access-smpgz\") pod \"cert-manager-cainjector-7d9f95dbf-9x9xp\" (UID: \"7897ec6d-8c3f-4f9b-93b2-489e21dc01af\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.719134 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7897ec6d-8c3f-4f9b-93b2-489e21dc01af-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-9x9xp\" (UID: \"7897ec6d-8c3f-4f9b-93b2-489e21dc01af\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.719218 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smpgz\" (UniqueName: \"kubernetes.io/projected/7897ec6d-8c3f-4f9b-93b2-489e21dc01af-kube-api-access-smpgz\") pod \"cert-manager-cainjector-7d9f95dbf-9x9xp\" (UID: \"7897ec6d-8c3f-4f9b-93b2-489e21dc01af\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.746284 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smpgz\" (UniqueName: \"kubernetes.io/projected/7897ec6d-8c3f-4f9b-93b2-489e21dc01af-kube-api-access-smpgz\") pod \"cert-manager-cainjector-7d9f95dbf-9x9xp\" (UID: \"7897ec6d-8c3f-4f9b-93b2-489e21dc01af\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.751845 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7897ec6d-8c3f-4f9b-93b2-489e21dc01af-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-9x9xp\" (UID: \"7897ec6d-8c3f-4f9b-93b2-489e21dc01af\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" Oct 07 22:24:48 crc kubenswrapper[4871]: I1007 22:24:48.911223 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" Oct 07 22:24:49 crc kubenswrapper[4871]: I1007 22:24:49.365053 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp"] Oct 07 22:24:50 crc kubenswrapper[4871]: I1007 22:24:50.097383 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" event={"ID":"7897ec6d-8c3f-4f9b-93b2-489e21dc01af","Type":"ContainerStarted","Data":"3b0e0783996f2bd0a180305036b896a77bae166d5ccce94b6f50474653b1afd3"} Oct 07 22:24:52 crc kubenswrapper[4871]: I1007 22:24:52.113032 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-zls94" event={"ID":"90172f12-e077-4e3a-93a0-115a93588eac","Type":"ContainerStarted","Data":"48cdbd1e2bf88cc6c47113f39d1f05ac304a6537dd2687a1da8fb20b431f0b6f"} Oct 07 22:24:52 crc kubenswrapper[4871]: I1007 22:24:52.113656 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:24:52 crc kubenswrapper[4871]: I1007 22:24:52.115215 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" event={"ID":"7897ec6d-8c3f-4f9b-93b2-489e21dc01af","Type":"ContainerStarted","Data":"4c4e7fa6684dd5be56c0fa350e4acdb8552d077aabc876c8e0de6b1831f4bde8"} Oct 07 22:24:52 crc kubenswrapper[4871]: I1007 22:24:52.137400 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-zls94" podStartSLOduration=1.340262028 podStartE2EDuration="7.137375073s" podCreationTimestamp="2025-10-07 22:24:45 +0000 UTC" firstStartedPulling="2025-10-07 22:24:46.016577447 +0000 UTC m=+959.819275550" lastFinishedPulling="2025-10-07 22:24:51.813690522 +0000 UTC m=+965.616388595" observedRunningTime="2025-10-07 22:24:52.134141768 +0000 UTC m=+965.936839851" watchObservedRunningTime="2025-10-07 22:24:52.137375073 +0000 UTC m=+965.940073156" Oct 07 22:24:52 crc kubenswrapper[4871]: I1007 22:24:52.155073 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-9x9xp" podStartSLOduration=1.723578276 podStartE2EDuration="4.155049037s" podCreationTimestamp="2025-10-07 22:24:48 +0000 UTC" firstStartedPulling="2025-10-07 22:24:49.380917237 +0000 UTC m=+963.183615310" lastFinishedPulling="2025-10-07 22:24:51.812387998 +0000 UTC m=+965.615086071" observedRunningTime="2025-10-07 22:24:52.152550612 +0000 UTC m=+965.955248725" watchObservedRunningTime="2025-10-07 22:24:52.155049037 +0000 UTC m=+965.957747120" Oct 07 22:25:00 crc kubenswrapper[4871]: I1007 22:25:00.504258 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-zls94" Oct 07 22:25:04 crc kubenswrapper[4871]: I1007 22:25:04.920564 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-p5q48"] Oct 07 22:25:04 crc kubenswrapper[4871]: I1007 22:25:04.922199 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" Oct 07 22:25:04 crc kubenswrapper[4871]: I1007 22:25:04.935184 4871 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-624qf" Oct 07 22:25:04 crc kubenswrapper[4871]: I1007 22:25:04.935857 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-p5q48"] Oct 07 22:25:05 crc kubenswrapper[4871]: I1007 22:25:05.103583 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brvmx\" (UniqueName: \"kubernetes.io/projected/4a764d87-e672-42a7-b88c-d1485bfd7c19-kube-api-access-brvmx\") pod \"cert-manager-7d4cc89fcb-p5q48\" (UID: \"4a764d87-e672-42a7-b88c-d1485bfd7c19\") " pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" Oct 07 22:25:05 crc kubenswrapper[4871]: I1007 22:25:05.103718 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a764d87-e672-42a7-b88c-d1485bfd7c19-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-p5q48\" (UID: \"4a764d87-e672-42a7-b88c-d1485bfd7c19\") " pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" Oct 07 22:25:05 crc kubenswrapper[4871]: I1007 22:25:05.205850 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brvmx\" (UniqueName: \"kubernetes.io/projected/4a764d87-e672-42a7-b88c-d1485bfd7c19-kube-api-access-brvmx\") pod \"cert-manager-7d4cc89fcb-p5q48\" (UID: \"4a764d87-e672-42a7-b88c-d1485bfd7c19\") " pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" Oct 07 22:25:05 crc kubenswrapper[4871]: I1007 22:25:05.205973 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a764d87-e672-42a7-b88c-d1485bfd7c19-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-p5q48\" (UID: \"4a764d87-e672-42a7-b88c-d1485bfd7c19\") " pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" Oct 07 22:25:05 crc kubenswrapper[4871]: I1007 22:25:05.236958 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a764d87-e672-42a7-b88c-d1485bfd7c19-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-p5q48\" (UID: \"4a764d87-e672-42a7-b88c-d1485bfd7c19\") " pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" Oct 07 22:25:05 crc kubenswrapper[4871]: I1007 22:25:05.241216 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brvmx\" (UniqueName: \"kubernetes.io/projected/4a764d87-e672-42a7-b88c-d1485bfd7c19-kube-api-access-brvmx\") pod \"cert-manager-7d4cc89fcb-p5q48\" (UID: \"4a764d87-e672-42a7-b88c-d1485bfd7c19\") " pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" Oct 07 22:25:05 crc kubenswrapper[4871]: I1007 22:25:05.261262 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" Oct 07 22:25:05 crc kubenswrapper[4871]: I1007 22:25:05.520414 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-p5q48"] Oct 07 22:25:05 crc kubenswrapper[4871]: W1007 22:25:05.530125 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a764d87_e672_42a7_b88c_d1485bfd7c19.slice/crio-0dd90f35c9057f643899b0f10b361da64a3fea527854d0d13f06b68e3d5d7cc3 WatchSource:0}: Error finding container 0dd90f35c9057f643899b0f10b361da64a3fea527854d0d13f06b68e3d5d7cc3: Status 404 returned error can't find the container with id 0dd90f35c9057f643899b0f10b361da64a3fea527854d0d13f06b68e3d5d7cc3 Oct 07 22:25:06 crc kubenswrapper[4871]: I1007 22:25:06.243089 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" event={"ID":"4a764d87-e672-42a7-b88c-d1485bfd7c19","Type":"ContainerStarted","Data":"717816f6ca86c8ccfe43acdfbfb765cabf9c0a74f379e7212a71040bf538517c"} Oct 07 22:25:06 crc kubenswrapper[4871]: I1007 22:25:06.245958 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" event={"ID":"4a764d87-e672-42a7-b88c-d1485bfd7c19","Type":"ContainerStarted","Data":"0dd90f35c9057f643899b0f10b361da64a3fea527854d0d13f06b68e3d5d7cc3"} Oct 07 22:25:06 crc kubenswrapper[4871]: I1007 22:25:06.269295 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-p5q48" podStartSLOduration=2.269260704 podStartE2EDuration="2.269260704s" podCreationTimestamp="2025-10-07 22:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:25:06.267113478 +0000 UTC m=+980.069811591" watchObservedRunningTime="2025-10-07 22:25:06.269260704 +0000 UTC m=+980.071958817" Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.253273 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-lfbnq"] Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.255181 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lfbnq" Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.262723 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.267145 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.270272 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-rf98h" Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.274334 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lfbnq"] Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.362786 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwx72\" (UniqueName: \"kubernetes.io/projected/f17098b5-641f-46cc-a699-525cbd8c6740-kube-api-access-pwx72\") pod \"openstack-operator-index-lfbnq\" (UID: \"f17098b5-641f-46cc-a699-525cbd8c6740\") " pod="openstack-operators/openstack-operator-index-lfbnq" Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.464195 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwx72\" (UniqueName: \"kubernetes.io/projected/f17098b5-641f-46cc-a699-525cbd8c6740-kube-api-access-pwx72\") pod \"openstack-operator-index-lfbnq\" (UID: \"f17098b5-641f-46cc-a699-525cbd8c6740\") " pod="openstack-operators/openstack-operator-index-lfbnq" Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.485334 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwx72\" (UniqueName: \"kubernetes.io/projected/f17098b5-641f-46cc-a699-525cbd8c6740-kube-api-access-pwx72\") pod \"openstack-operator-index-lfbnq\" (UID: \"f17098b5-641f-46cc-a699-525cbd8c6740\") " pod="openstack-operators/openstack-operator-index-lfbnq" Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.625723 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lfbnq" Oct 07 22:25:14 crc kubenswrapper[4871]: I1007 22:25:14.891249 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lfbnq"] Oct 07 22:25:14 crc kubenswrapper[4871]: W1007 22:25:14.898503 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf17098b5_641f_46cc_a699_525cbd8c6740.slice/crio-e607fbf4fc36ad9345f08c7189749873f068c5612078f2709ee19e52a2465025 WatchSource:0}: Error finding container e607fbf4fc36ad9345f08c7189749873f068c5612078f2709ee19e52a2465025: Status 404 returned error can't find the container with id e607fbf4fc36ad9345f08c7189749873f068c5612078f2709ee19e52a2465025 Oct 07 22:25:15 crc kubenswrapper[4871]: I1007 22:25:15.327410 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lfbnq" event={"ID":"f17098b5-641f-46cc-a699-525cbd8c6740","Type":"ContainerStarted","Data":"e607fbf4fc36ad9345f08c7189749873f068c5612078f2709ee19e52a2465025"} Oct 07 22:25:17 crc kubenswrapper[4871]: I1007 22:25:17.350677 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lfbnq" event={"ID":"f17098b5-641f-46cc-a699-525cbd8c6740","Type":"ContainerStarted","Data":"d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799"} Oct 07 22:25:17 crc kubenswrapper[4871]: I1007 22:25:17.380707 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-lfbnq" podStartSLOduration=1.5595373559999999 podStartE2EDuration="3.38067319s" podCreationTimestamp="2025-10-07 22:25:14 +0000 UTC" firstStartedPulling="2025-10-07 22:25:14.902288815 +0000 UTC m=+988.704986888" lastFinishedPulling="2025-10-07 22:25:16.723424659 +0000 UTC m=+990.526122722" observedRunningTime="2025-10-07 22:25:17.375606157 +0000 UTC m=+991.178304270" watchObservedRunningTime="2025-10-07 22:25:17.38067319 +0000 UTC m=+991.183371303" Oct 07 22:25:17 crc kubenswrapper[4871]: I1007 22:25:17.624594 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lfbnq"] Oct 07 22:25:18 crc kubenswrapper[4871]: I1007 22:25:18.241938 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-cgtsg"] Oct 07 22:25:18 crc kubenswrapper[4871]: I1007 22:25:18.243623 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cgtsg" Oct 07 22:25:18 crc kubenswrapper[4871]: I1007 22:25:18.253686 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-cgtsg"] Oct 07 22:25:18 crc kubenswrapper[4871]: I1007 22:25:18.433466 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crksx\" (UniqueName: \"kubernetes.io/projected/e8cf523f-4104-44ad-b446-86ad5318cc9f-kube-api-access-crksx\") pod \"openstack-operator-index-cgtsg\" (UID: \"e8cf523f-4104-44ad-b446-86ad5318cc9f\") " pod="openstack-operators/openstack-operator-index-cgtsg" Oct 07 22:25:18 crc kubenswrapper[4871]: I1007 22:25:18.536580 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crksx\" (UniqueName: \"kubernetes.io/projected/e8cf523f-4104-44ad-b446-86ad5318cc9f-kube-api-access-crksx\") pod \"openstack-operator-index-cgtsg\" (UID: \"e8cf523f-4104-44ad-b446-86ad5318cc9f\") " pod="openstack-operators/openstack-operator-index-cgtsg" Oct 07 22:25:18 crc kubenswrapper[4871]: I1007 22:25:18.585956 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crksx\" (UniqueName: \"kubernetes.io/projected/e8cf523f-4104-44ad-b446-86ad5318cc9f-kube-api-access-crksx\") pod \"openstack-operator-index-cgtsg\" (UID: \"e8cf523f-4104-44ad-b446-86ad5318cc9f\") " pod="openstack-operators/openstack-operator-index-cgtsg" Oct 07 22:25:18 crc kubenswrapper[4871]: I1007 22:25:18.868655 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cgtsg" Oct 07 22:25:19 crc kubenswrapper[4871]: I1007 22:25:19.104169 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-cgtsg"] Oct 07 22:25:19 crc kubenswrapper[4871]: I1007 22:25:19.374739 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cgtsg" event={"ID":"e8cf523f-4104-44ad-b446-86ad5318cc9f","Type":"ContainerStarted","Data":"519c55f41da2ecdc7aed9805e72f46a265139b72c3cb4507dfd7236d2042f440"} Oct 07 22:25:19 crc kubenswrapper[4871]: I1007 22:25:19.374933 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-lfbnq" podUID="f17098b5-641f-46cc-a699-525cbd8c6740" containerName="registry-server" containerID="cri-o://d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799" gracePeriod=2 Oct 07 22:25:19 crc kubenswrapper[4871]: I1007 22:25:19.768211 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lfbnq" Oct 07 22:25:19 crc kubenswrapper[4871]: I1007 22:25:19.860702 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwx72\" (UniqueName: \"kubernetes.io/projected/f17098b5-641f-46cc-a699-525cbd8c6740-kube-api-access-pwx72\") pod \"f17098b5-641f-46cc-a699-525cbd8c6740\" (UID: \"f17098b5-641f-46cc-a699-525cbd8c6740\") " Oct 07 22:25:19 crc kubenswrapper[4871]: I1007 22:25:19.869041 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f17098b5-641f-46cc-a699-525cbd8c6740-kube-api-access-pwx72" (OuterVolumeSpecName: "kube-api-access-pwx72") pod "f17098b5-641f-46cc-a699-525cbd8c6740" (UID: "f17098b5-641f-46cc-a699-525cbd8c6740"). InnerVolumeSpecName "kube-api-access-pwx72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:25:19 crc kubenswrapper[4871]: I1007 22:25:19.962298 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwx72\" (UniqueName: \"kubernetes.io/projected/f17098b5-641f-46cc-a699-525cbd8c6740-kube-api-access-pwx72\") on node \"crc\" DevicePath \"\"" Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.384197 4871 generic.go:334] "Generic (PLEG): container finished" podID="f17098b5-641f-46cc-a699-525cbd8c6740" containerID="d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799" exitCode=0 Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.384280 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lfbnq" Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.384286 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lfbnq" event={"ID":"f17098b5-641f-46cc-a699-525cbd8c6740","Type":"ContainerDied","Data":"d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799"} Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.384999 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lfbnq" event={"ID":"f17098b5-641f-46cc-a699-525cbd8c6740","Type":"ContainerDied","Data":"e607fbf4fc36ad9345f08c7189749873f068c5612078f2709ee19e52a2465025"} Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.385038 4871 scope.go:117] "RemoveContainer" containerID="d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799" Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.408368 4871 scope.go:117] "RemoveContainer" containerID="d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799" Oct 07 22:25:20 crc kubenswrapper[4871]: E1007 22:25:20.408996 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799\": container with ID starting with d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799 not found: ID does not exist" containerID="d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799" Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.409104 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799"} err="failed to get container status \"d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799\": rpc error: code = NotFound desc = could not find container \"d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799\": container with ID starting with d153fffbe0bddf147b0f4122a1383bbff1fba5f4575ad0059567ebd66c4f3799 not found: ID does not exist" Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.437827 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lfbnq"] Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.437937 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-lfbnq"] Oct 07 22:25:20 crc kubenswrapper[4871]: I1007 22:25:20.995623 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f17098b5-641f-46cc-a699-525cbd8c6740" path="/var/lib/kubelet/pods/f17098b5-641f-46cc-a699-525cbd8c6740/volumes" Oct 07 22:25:21 crc kubenswrapper[4871]: I1007 22:25:21.395112 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cgtsg" event={"ID":"e8cf523f-4104-44ad-b446-86ad5318cc9f","Type":"ContainerStarted","Data":"94a69e23768d4141a9fdaf47ecbe14433282e839bc340fb08d63681262bc963d"} Oct 07 22:25:28 crc kubenswrapper[4871]: I1007 22:25:28.869351 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-cgtsg" Oct 07 22:25:28 crc kubenswrapper[4871]: I1007 22:25:28.870362 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-cgtsg" Oct 07 22:25:28 crc kubenswrapper[4871]: I1007 22:25:28.916907 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-cgtsg" Oct 07 22:25:28 crc kubenswrapper[4871]: I1007 22:25:28.943703 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-cgtsg" podStartSLOduration=9.602561384 podStartE2EDuration="10.943656366s" podCreationTimestamp="2025-10-07 22:25:18 +0000 UTC" firstStartedPulling="2025-10-07 22:25:19.112983818 +0000 UTC m=+992.915681891" lastFinishedPulling="2025-10-07 22:25:20.4540788 +0000 UTC m=+994.256776873" observedRunningTime="2025-10-07 22:25:21.413465596 +0000 UTC m=+995.216163669" watchObservedRunningTime="2025-10-07 22:25:28.943656366 +0000 UTC m=+1002.746354439" Oct 07 22:25:29 crc kubenswrapper[4871]: I1007 22:25:29.499358 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-cgtsg" Oct 07 22:25:36 crc kubenswrapper[4871]: I1007 22:25:36.895107 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp"] Oct 07 22:25:36 crc kubenswrapper[4871]: E1007 22:25:36.895879 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17098b5-641f-46cc-a699-525cbd8c6740" containerName="registry-server" Oct 07 22:25:36 crc kubenswrapper[4871]: I1007 22:25:36.895894 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17098b5-641f-46cc-a699-525cbd8c6740" containerName="registry-server" Oct 07 22:25:36 crc kubenswrapper[4871]: I1007 22:25:36.896057 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17098b5-641f-46cc-a699-525cbd8c6740" containerName="registry-server" Oct 07 22:25:36 crc kubenswrapper[4871]: I1007 22:25:36.897070 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:36 crc kubenswrapper[4871]: I1007 22:25:36.899937 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dqj5l" Oct 07 22:25:36 crc kubenswrapper[4871]: I1007 22:25:36.920858 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp"] Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.040414 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24kwm\" (UniqueName: \"kubernetes.io/projected/fe082719-5871-4673-b838-cf19f5b6cd45-kube-api-access-24kwm\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.040484 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-util\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.040548 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-bundle\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.142283 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24kwm\" (UniqueName: \"kubernetes.io/projected/fe082719-5871-4673-b838-cf19f5b6cd45-kube-api-access-24kwm\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.142372 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-util\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.142512 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-bundle\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.143589 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-util\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.143609 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-bundle\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.170958 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24kwm\" (UniqueName: \"kubernetes.io/projected/fe082719-5871-4673-b838-cf19f5b6cd45-kube-api-access-24kwm\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.226285 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.470263 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp"] Oct 07 22:25:37 crc kubenswrapper[4871]: W1007 22:25:37.479451 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe082719_5871_4673_b838_cf19f5b6cd45.slice/crio-aa3f42f0459dd74b5620efaf3edea66081f1f286fd8f7db9a9aeb8bf2fad71fd WatchSource:0}: Error finding container aa3f42f0459dd74b5620efaf3edea66081f1f286fd8f7db9a9aeb8bf2fad71fd: Status 404 returned error can't find the container with id aa3f42f0459dd74b5620efaf3edea66081f1f286fd8f7db9a9aeb8bf2fad71fd Oct 07 22:25:37 crc kubenswrapper[4871]: I1007 22:25:37.525160 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" event={"ID":"fe082719-5871-4673-b838-cf19f5b6cd45","Type":"ContainerStarted","Data":"aa3f42f0459dd74b5620efaf3edea66081f1f286fd8f7db9a9aeb8bf2fad71fd"} Oct 07 22:25:38 crc kubenswrapper[4871]: I1007 22:25:38.537410 4871 generic.go:334] "Generic (PLEG): container finished" podID="fe082719-5871-4673-b838-cf19f5b6cd45" containerID="ed75059ca25c107dc36d82982639d456cd9039ea6d25168765d9ff855d6627f1" exitCode=0 Oct 07 22:25:38 crc kubenswrapper[4871]: I1007 22:25:38.537487 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" event={"ID":"fe082719-5871-4673-b838-cf19f5b6cd45","Type":"ContainerDied","Data":"ed75059ca25c107dc36d82982639d456cd9039ea6d25168765d9ff855d6627f1"} Oct 07 22:25:40 crc kubenswrapper[4871]: I1007 22:25:40.560814 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" event={"ID":"fe082719-5871-4673-b838-cf19f5b6cd45","Type":"ContainerStarted","Data":"249cdbd9e9e087dd86e8734359d8b61e27d97a190de219eac5ee0fb1f8b0b94d"} Oct 07 22:25:41 crc kubenswrapper[4871]: I1007 22:25:41.573153 4871 generic.go:334] "Generic (PLEG): container finished" podID="fe082719-5871-4673-b838-cf19f5b6cd45" containerID="249cdbd9e9e087dd86e8734359d8b61e27d97a190de219eac5ee0fb1f8b0b94d" exitCode=0 Oct 07 22:25:41 crc kubenswrapper[4871]: I1007 22:25:41.573240 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" event={"ID":"fe082719-5871-4673-b838-cf19f5b6cd45","Type":"ContainerDied","Data":"249cdbd9e9e087dd86e8734359d8b61e27d97a190de219eac5ee0fb1f8b0b94d"} Oct 07 22:25:42 crc kubenswrapper[4871]: I1007 22:25:42.585141 4871 generic.go:334] "Generic (PLEG): container finished" podID="fe082719-5871-4673-b838-cf19f5b6cd45" containerID="ad5349602f087fb45b3e9e83c2935c65d3fa3517927b030416e99602fc5fdbca" exitCode=0 Oct 07 22:25:42 crc kubenswrapper[4871]: I1007 22:25:42.585236 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" event={"ID":"fe082719-5871-4673-b838-cf19f5b6cd45","Type":"ContainerDied","Data":"ad5349602f087fb45b3e9e83c2935c65d3fa3517927b030416e99602fc5fdbca"} Oct 07 22:25:43 crc kubenswrapper[4871]: I1007 22:25:43.905025 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.050786 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24kwm\" (UniqueName: \"kubernetes.io/projected/fe082719-5871-4673-b838-cf19f5b6cd45-kube-api-access-24kwm\") pod \"fe082719-5871-4673-b838-cf19f5b6cd45\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.050909 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-bundle\") pod \"fe082719-5871-4673-b838-cf19f5b6cd45\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.051037 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-util\") pod \"fe082719-5871-4673-b838-cf19f5b6cd45\" (UID: \"fe082719-5871-4673-b838-cf19f5b6cd45\") " Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.051898 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-bundle" (OuterVolumeSpecName: "bundle") pod "fe082719-5871-4673-b838-cf19f5b6cd45" (UID: "fe082719-5871-4673-b838-cf19f5b6cd45"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.058105 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe082719-5871-4673-b838-cf19f5b6cd45-kube-api-access-24kwm" (OuterVolumeSpecName: "kube-api-access-24kwm") pod "fe082719-5871-4673-b838-cf19f5b6cd45" (UID: "fe082719-5871-4673-b838-cf19f5b6cd45"). InnerVolumeSpecName "kube-api-access-24kwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.066443 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-util" (OuterVolumeSpecName: "util") pod "fe082719-5871-4673-b838-cf19f5b6cd45" (UID: "fe082719-5871-4673-b838-cf19f5b6cd45"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.153269 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24kwm\" (UniqueName: \"kubernetes.io/projected/fe082719-5871-4673-b838-cf19f5b6cd45-kube-api-access-24kwm\") on node \"crc\" DevicePath \"\"" Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.153329 4871 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.153348 4871 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fe082719-5871-4673-b838-cf19f5b6cd45-util\") on node \"crc\" DevicePath \"\"" Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.602189 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" event={"ID":"fe082719-5871-4673-b838-cf19f5b6cd45","Type":"ContainerDied","Data":"aa3f42f0459dd74b5620efaf3edea66081f1f286fd8f7db9a9aeb8bf2fad71fd"} Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.602250 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa3f42f0459dd74b5620efaf3edea66081f1f286fd8f7db9a9aeb8bf2fad71fd" Oct 07 22:25:44 crc kubenswrapper[4871]: I1007 22:25:44.602656 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.361911 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s"] Oct 07 22:25:49 crc kubenswrapper[4871]: E1007 22:25:49.363013 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe082719-5871-4673-b838-cf19f5b6cd45" containerName="util" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.363030 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe082719-5871-4673-b838-cf19f5b6cd45" containerName="util" Oct 07 22:25:49 crc kubenswrapper[4871]: E1007 22:25:49.363047 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe082719-5871-4673-b838-cf19f5b6cd45" containerName="extract" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.363055 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe082719-5871-4673-b838-cf19f5b6cd45" containerName="extract" Oct 07 22:25:49 crc kubenswrapper[4871]: E1007 22:25:49.363072 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe082719-5871-4673-b838-cf19f5b6cd45" containerName="pull" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.363082 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe082719-5871-4673-b838-cf19f5b6cd45" containerName="pull" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.363222 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe082719-5871-4673-b838-cf19f5b6cd45" containerName="extract" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.364062 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.365850 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-pgt2m" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.411251 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s"] Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.535800 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fxrr\" (UniqueName: \"kubernetes.io/projected/6198e2f8-88e4-4ad8-bda0-aab47a10f07a-kube-api-access-9fxrr\") pod \"openstack-operator-controller-operator-b6d857f89-scc7s\" (UID: \"6198e2f8-88e4-4ad8-bda0-aab47a10f07a\") " pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.637507 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fxrr\" (UniqueName: \"kubernetes.io/projected/6198e2f8-88e4-4ad8-bda0-aab47a10f07a-kube-api-access-9fxrr\") pod \"openstack-operator-controller-operator-b6d857f89-scc7s\" (UID: \"6198e2f8-88e4-4ad8-bda0-aab47a10f07a\") " pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.664947 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fxrr\" (UniqueName: \"kubernetes.io/projected/6198e2f8-88e4-4ad8-bda0-aab47a10f07a-kube-api-access-9fxrr\") pod \"openstack-operator-controller-operator-b6d857f89-scc7s\" (UID: \"6198e2f8-88e4-4ad8-bda0-aab47a10f07a\") " pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" Oct 07 22:25:49 crc kubenswrapper[4871]: I1007 22:25:49.680304 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" Oct 07 22:25:50 crc kubenswrapper[4871]: I1007 22:25:50.141167 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s"] Oct 07 22:25:50 crc kubenswrapper[4871]: I1007 22:25:50.654210 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" event={"ID":"6198e2f8-88e4-4ad8-bda0-aab47a10f07a","Type":"ContainerStarted","Data":"f1b11ba49f03732d6c5cd1bafa901a0892c02cf01985c19085f5b3f57f423544"} Oct 07 22:25:55 crc kubenswrapper[4871]: I1007 22:25:55.690930 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" event={"ID":"6198e2f8-88e4-4ad8-bda0-aab47a10f07a","Type":"ContainerStarted","Data":"b5d9fff2eb4f0fbe4bdc3e16f2ae9ddd98156478ce88947575214eb9946b8624"} Oct 07 22:25:59 crc kubenswrapper[4871]: I1007 22:25:59.720590 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" event={"ID":"6198e2f8-88e4-4ad8-bda0-aab47a10f07a","Type":"ContainerStarted","Data":"455ca47789493f79f7d7b3e34cb11eca6ea0a6eff0d0b8e26eb65d667099a7b9"} Oct 07 22:25:59 crc kubenswrapper[4871]: I1007 22:25:59.721446 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" Oct 07 22:25:59 crc kubenswrapper[4871]: I1007 22:25:59.762671 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" podStartSLOduration=1.894072901 podStartE2EDuration="10.762641434s" podCreationTimestamp="2025-10-07 22:25:49 +0000 UTC" firstStartedPulling="2025-10-07 22:25:50.154177117 +0000 UTC m=+1023.956875230" lastFinishedPulling="2025-10-07 22:25:59.02274569 +0000 UTC m=+1032.825443763" observedRunningTime="2025-10-07 22:25:59.75715074 +0000 UTC m=+1033.559848873" watchObservedRunningTime="2025-10-07 22:25:59.762641434 +0000 UTC m=+1033.565339497" Oct 07 22:26:00 crc kubenswrapper[4871]: I1007 22:26:00.732328 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-scc7s" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.196060 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.198400 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.200572 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-48rdt" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.208139 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.209335 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.216252 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-ljdkp" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.217061 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.225124 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.226498 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.230502 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.233064 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-cqknx" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.240467 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.251817 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvk9m\" (UniqueName: \"kubernetes.io/projected/07b6d737-7f0c-494d-adfe-0a641261c344-kube-api-access-bvk9m\") pod \"barbican-operator-controller-manager-64f56ff694-fh5vz\" (UID: \"07b6d737-7f0c-494d-adfe-0a641261c344\") " pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.252182 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sc6f\" (UniqueName: \"kubernetes.io/projected/b58f786d-7c8a-4506-94d4-bf1fcf543871-kube-api-access-2sc6f\") pod \"cinder-operator-controller-manager-84bd8f6848-v6x97\" (UID: \"b58f786d-7c8a-4506-94d4-bf1fcf543871\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.252358 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p6gr\" (UniqueName: \"kubernetes.io/projected/e65e1c1c-1bbc-488e-821c-790bb85bcd55-kube-api-access-9p6gr\") pod \"designate-operator-controller-manager-58d86cd59d-5ndrn\" (UID: \"e65e1c1c-1bbc-488e-821c-790bb85bcd55\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.252503 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.253903 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.257021 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-sr8ds" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.295968 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.307027 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.308525 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.316561 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-tfmtm" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.323638 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.335135 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.336537 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.338511 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-4r4xn" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.365513 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sc6f\" (UniqueName: \"kubernetes.io/projected/b58f786d-7c8a-4506-94d4-bf1fcf543871-kube-api-access-2sc6f\") pod \"cinder-operator-controller-manager-84bd8f6848-v6x97\" (UID: \"b58f786d-7c8a-4506-94d4-bf1fcf543871\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.369521 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p6gr\" (UniqueName: \"kubernetes.io/projected/e65e1c1c-1bbc-488e-821c-790bb85bcd55-kube-api-access-9p6gr\") pod \"designate-operator-controller-manager-58d86cd59d-5ndrn\" (UID: \"e65e1c1c-1bbc-488e-821c-790bb85bcd55\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.369916 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvk9m\" (UniqueName: \"kubernetes.io/projected/07b6d737-7f0c-494d-adfe-0a641261c344-kube-api-access-bvk9m\") pod \"barbican-operator-controller-manager-64f56ff694-fh5vz\" (UID: \"07b6d737-7f0c-494d-adfe-0a641261c344\") " pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.433355 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvk9m\" (UniqueName: \"kubernetes.io/projected/07b6d737-7f0c-494d-adfe-0a641261c344-kube-api-access-bvk9m\") pod \"barbican-operator-controller-manager-64f56ff694-fh5vz\" (UID: \"07b6d737-7f0c-494d-adfe-0a641261c344\") " pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.436217 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sc6f\" (UniqueName: \"kubernetes.io/projected/b58f786d-7c8a-4506-94d4-bf1fcf543871-kube-api-access-2sc6f\") pod \"cinder-operator-controller-manager-84bd8f6848-v6x97\" (UID: \"b58f786d-7c8a-4506-94d4-bf1fcf543871\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.436231 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p6gr\" (UniqueName: \"kubernetes.io/projected/e65e1c1c-1bbc-488e-821c-790bb85bcd55-kube-api-access-9p6gr\") pod \"designate-operator-controller-manager-58d86cd59d-5ndrn\" (UID: \"e65e1c1c-1bbc-488e-821c-790bb85bcd55\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.442902 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.444780 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.448891 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.453869 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.455331 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.455658 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.458593 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-6xt5n" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.460310 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.462676 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-qdmtf" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.466087 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.467331 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.478205 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkh4x\" (UniqueName: \"kubernetes.io/projected/e85a9c1c-2748-4ae7-a4fa-cfb555348feb-kube-api-access-kkh4x\") pod \"glance-operator-controller-manager-fd648f65-fsrt7\" (UID: \"e85a9c1c-2748-4ae7-a4fa-cfb555348feb\") " pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.478559 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.478673 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-kcwps" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.478849 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drg7q\" (UniqueName: \"kubernetes.io/projected/63b31dae-e86d-4735-96b8-125a9a066a3e-kube-api-access-drg7q\") pod \"ironic-operator-controller-manager-5467f8988c-ftlvg\" (UID: \"63b31dae-e86d-4735-96b8-125a9a066a3e\") " pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.478878 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwsbr\" (UniqueName: \"kubernetes.io/projected/b3a4e444-6508-4ae3-8b40-3f238f252ed2-kube-api-access-bwsbr\") pod \"keystone-operator-controller-manager-5b84cc7657-fwjnb\" (UID: \"b3a4e444-6508-4ae3-8b40-3f238f252ed2\") " pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.478931 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzmnh\" (UniqueName: \"kubernetes.io/projected/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-kube-api-access-zzmnh\") pod \"infra-operator-controller-manager-84788b6bc5-c6bxk\" (UID: \"e08fd5ec-5b08-4ad9-b298-886bda95d3d9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.478978 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flk5d\" (UniqueName: \"kubernetes.io/projected/8acf9fc2-6307-4d04-921c-cdfc851f8c52-kube-api-access-flk5d\") pod \"heat-operator-controller-manager-7ccfc8cf49-lg6tk\" (UID: \"8acf9fc2-6307-4d04-921c-cdfc851f8c52\") " pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.478995 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-cert\") pod \"infra-operator-controller-manager-84788b6bc5-c6bxk\" (UID: \"e08fd5ec-5b08-4ad9-b298-886bda95d3d9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.479012 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqqqf\" (UniqueName: \"kubernetes.io/projected/6916e675-a275-4e0f-b43d-c40a8973f13f-kube-api-access-mqqqf\") pod \"horizon-operator-controller-manager-5b477879bc-jd9ql\" (UID: \"6916e675-a275-4e0f-b43d-c40a8973f13f\") " pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.512882 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.514352 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.517639 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-69wv8" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.521402 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.523043 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.524819 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.528271 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-xttxh" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.536452 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.536576 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.541658 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.550664 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.552957 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.565463 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.565588 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.575619 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.577455 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-kdb2l" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.580671 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkh4x\" (UniqueName: \"kubernetes.io/projected/e85a9c1c-2748-4ae7-a4fa-cfb555348feb-kube-api-access-kkh4x\") pod \"glance-operator-controller-manager-fd648f65-fsrt7\" (UID: \"e85a9c1c-2748-4ae7-a4fa-cfb555348feb\") " pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.580755 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drg7q\" (UniqueName: \"kubernetes.io/projected/63b31dae-e86d-4735-96b8-125a9a066a3e-kube-api-access-drg7q\") pod \"ironic-operator-controller-manager-5467f8988c-ftlvg\" (UID: \"63b31dae-e86d-4735-96b8-125a9a066a3e\") " pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.580782 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwsbr\" (UniqueName: \"kubernetes.io/projected/b3a4e444-6508-4ae3-8b40-3f238f252ed2-kube-api-access-bwsbr\") pod \"keystone-operator-controller-manager-5b84cc7657-fwjnb\" (UID: \"b3a4e444-6508-4ae3-8b40-3f238f252ed2\") " pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.580826 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzmnh\" (UniqueName: \"kubernetes.io/projected/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-kube-api-access-zzmnh\") pod \"infra-operator-controller-manager-84788b6bc5-c6bxk\" (UID: \"e08fd5ec-5b08-4ad9-b298-886bda95d3d9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.580852 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flk5d\" (UniqueName: \"kubernetes.io/projected/8acf9fc2-6307-4d04-921c-cdfc851f8c52-kube-api-access-flk5d\") pod \"heat-operator-controller-manager-7ccfc8cf49-lg6tk\" (UID: \"8acf9fc2-6307-4d04-921c-cdfc851f8c52\") " pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.580884 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-cert\") pod \"infra-operator-controller-manager-84788b6bc5-c6bxk\" (UID: \"e08fd5ec-5b08-4ad9-b298-886bda95d3d9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.580909 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqqqf\" (UniqueName: \"kubernetes.io/projected/6916e675-a275-4e0f-b43d-c40a8973f13f-kube-api-access-mqqqf\") pod \"horizon-operator-controller-manager-5b477879bc-jd9ql\" (UID: \"6916e675-a275-4e0f-b43d-c40a8973f13f\") " pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" Oct 07 22:26:33 crc kubenswrapper[4871]: E1007 22:26:33.582852 4871 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 07 22:26:33 crc kubenswrapper[4871]: E1007 22:26:33.582993 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-cert podName:e08fd5ec-5b08-4ad9-b298-886bda95d3d9 nodeName:}" failed. No retries permitted until 2025-10-07 22:26:34.082969946 +0000 UTC m=+1067.885668019 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-cert") pod "infra-operator-controller-manager-84788b6bc5-c6bxk" (UID: "e08fd5ec-5b08-4ad9-b298-886bda95d3d9") : secret "infra-operator-webhook-server-cert" not found Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.596709 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.596747 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.596838 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.605171 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-fgfjs" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.615729 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drg7q\" (UniqueName: \"kubernetes.io/projected/63b31dae-e86d-4735-96b8-125a9a066a3e-kube-api-access-drg7q\") pod \"ironic-operator-controller-manager-5467f8988c-ftlvg\" (UID: \"63b31dae-e86d-4735-96b8-125a9a066a3e\") " pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.616196 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.617255 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.620151 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkh4x\" (UniqueName: \"kubernetes.io/projected/e85a9c1c-2748-4ae7-a4fa-cfb555348feb-kube-api-access-kkh4x\") pod \"glance-operator-controller-manager-fd648f65-fsrt7\" (UID: \"e85a9c1c-2748-4ae7-a4fa-cfb555348feb\") " pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.620496 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-645tw" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.620744 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqqqf\" (UniqueName: \"kubernetes.io/projected/6916e675-a275-4e0f-b43d-c40a8973f13f-kube-api-access-mqqqf\") pod \"horizon-operator-controller-manager-5b477879bc-jd9ql\" (UID: \"6916e675-a275-4e0f-b43d-c40a8973f13f\") " pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.622972 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwsbr\" (UniqueName: \"kubernetes.io/projected/b3a4e444-6508-4ae3-8b40-3f238f252ed2-kube-api-access-bwsbr\") pod \"keystone-operator-controller-manager-5b84cc7657-fwjnb\" (UID: \"b3a4e444-6508-4ae3-8b40-3f238f252ed2\") " pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.628459 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flk5d\" (UniqueName: \"kubernetes.io/projected/8acf9fc2-6307-4d04-921c-cdfc851f8c52-kube-api-access-flk5d\") pod \"heat-operator-controller-manager-7ccfc8cf49-lg6tk\" (UID: \"8acf9fc2-6307-4d04-921c-cdfc851f8c52\") " pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.631934 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzmnh\" (UniqueName: \"kubernetes.io/projected/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-kube-api-access-zzmnh\") pod \"infra-operator-controller-manager-84788b6bc5-c6bxk\" (UID: \"e08fd5ec-5b08-4ad9-b298-886bda95d3d9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.638137 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.659232 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.665109 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.666760 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.670930 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.673083 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.673984 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-29db4" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.678296 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.678650 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-fz7t9" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.678775 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.686098 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66hrr\" (UniqueName: \"kubernetes.io/projected/36980ce9-53e4-401a-b093-bd1ba933ae83-kube-api-access-66hrr\") pod \"nova-operator-controller-manager-6c9b57c67-hk9db\" (UID: \"36980ce9-53e4-401a-b093-bd1ba933ae83\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.686189 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f919f707-59de-4f00-a9cf-177ed5696694-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686wj76g\" (UID: \"f919f707-59de-4f00-a9cf-177ed5696694\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.686246 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9gm2\" (UniqueName: \"kubernetes.io/projected/cda39518-2b25-43e2-8afa-b0ff0ec18ff7-kube-api-access-x9gm2\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-8bw69\" (UID: \"cda39518-2b25-43e2-8afa-b0ff0ec18ff7\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.686312 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhgxj\" (UniqueName: \"kubernetes.io/projected/f919f707-59de-4f00-a9cf-177ed5696694-kube-api-access-fhgxj\") pod \"openstack-baremetal-operator-controller-manager-6875c66686wj76g\" (UID: \"f919f707-59de-4f00-a9cf-177ed5696694\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.686360 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfbsw\" (UniqueName: \"kubernetes.io/projected/9772f309-5c98-45d7-a404-b14c2ab199e6-kube-api-access-rfbsw\") pod \"octavia-operator-controller-manager-69f59f9d8-qj54s\" (UID: \"9772f309-5c98-45d7-a404-b14c2ab199e6\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.686467 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txwt8\" (UniqueName: \"kubernetes.io/projected/fc43555d-cedf-469e-af48-b2f3cce766c0-kube-api-access-txwt8\") pod \"neutron-operator-controller-manager-69b956fbf6-czwls\" (UID: \"fc43555d-cedf-469e-af48-b2f3cce766c0\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.686554 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4b7d\" (UniqueName: \"kubernetes.io/projected/0c7c4f04-05a9-40a5-9cb1-9320e33444b1-kube-api-access-w4b7d\") pod \"manila-operator-controller-manager-7cb48dbc-rtfnp\" (UID: \"0c7c4f04-05a9-40a5-9cb1-9320e33444b1\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.686591 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px247\" (UniqueName: \"kubernetes.io/projected/f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6-kube-api-access-px247\") pod \"ovn-operator-controller-manager-54d485fd9-c5989\" (UID: \"f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6\") " pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.688012 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.692180 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.703262 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.707337 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.707690 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.708532 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.710623 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-9cl5l" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.710896 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-j7njr" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.711000 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-nz284"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.712421 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.713904 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-tvwjr" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.714118 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.717688 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.726286 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-nz284"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.754063 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.777499 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.781678 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-lr8tm" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.789894 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px247\" (UniqueName: \"kubernetes.io/projected/f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6-kube-api-access-px247\") pod \"ovn-operator-controller-manager-54d485fd9-c5989\" (UID: \"f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6\") " pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.789943 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66hrr\" (UniqueName: \"kubernetes.io/projected/36980ce9-53e4-401a-b093-bd1ba933ae83-kube-api-access-66hrr\") pod \"nova-operator-controller-manager-6c9b57c67-hk9db\" (UID: \"36980ce9-53e4-401a-b093-bd1ba933ae83\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.789988 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f919f707-59de-4f00-a9cf-177ed5696694-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686wj76g\" (UID: \"f919f707-59de-4f00-a9cf-177ed5696694\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.790031 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9gm2\" (UniqueName: \"kubernetes.io/projected/cda39518-2b25-43e2-8afa-b0ff0ec18ff7-kube-api-access-x9gm2\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-8bw69\" (UID: \"cda39518-2b25-43e2-8afa-b0ff0ec18ff7\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.790074 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhgxj\" (UniqueName: \"kubernetes.io/projected/f919f707-59de-4f00-a9cf-177ed5696694-kube-api-access-fhgxj\") pod \"openstack-baremetal-operator-controller-manager-6875c66686wj76g\" (UID: \"f919f707-59de-4f00-a9cf-177ed5696694\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.790103 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfbsw\" (UniqueName: \"kubernetes.io/projected/9772f309-5c98-45d7-a404-b14c2ab199e6-kube-api-access-rfbsw\") pod \"octavia-operator-controller-manager-69f59f9d8-qj54s\" (UID: \"9772f309-5c98-45d7-a404-b14c2ab199e6\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.790124 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txwt8\" (UniqueName: \"kubernetes.io/projected/fc43555d-cedf-469e-af48-b2f3cce766c0-kube-api-access-txwt8\") pod \"neutron-operator-controller-manager-69b956fbf6-czwls\" (UID: \"fc43555d-cedf-469e-af48-b2f3cce766c0\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.790168 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4b7d\" (UniqueName: \"kubernetes.io/projected/0c7c4f04-05a9-40a5-9cb1-9320e33444b1-kube-api-access-w4b7d\") pod \"manila-operator-controller-manager-7cb48dbc-rtfnp\" (UID: \"0c7c4f04-05a9-40a5-9cb1-9320e33444b1\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" Oct 07 22:26:33 crc kubenswrapper[4871]: E1007 22:26:33.794009 4871 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 22:26:33 crc kubenswrapper[4871]: E1007 22:26:33.794212 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f919f707-59de-4f00-a9cf-177ed5696694-cert podName:f919f707-59de-4f00-a9cf-177ed5696694 nodeName:}" failed. No retries permitted until 2025-10-07 22:26:34.294190073 +0000 UTC m=+1068.096888146 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f919f707-59de-4f00-a9cf-177ed5696694-cert") pod "openstack-baremetal-operator-controller-manager-6875c66686wj76g" (UID: "f919f707-59de-4f00-a9cf-177ed5696694") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.830298 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.835429 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhgxj\" (UniqueName: \"kubernetes.io/projected/f919f707-59de-4f00-a9cf-177ed5696694-kube-api-access-fhgxj\") pod \"openstack-baremetal-operator-controller-manager-6875c66686wj76g\" (UID: \"f919f707-59de-4f00-a9cf-177ed5696694\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.836965 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.837359 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66hrr\" (UniqueName: \"kubernetes.io/projected/36980ce9-53e4-401a-b093-bd1ba933ae83-kube-api-access-66hrr\") pod \"nova-operator-controller-manager-6c9b57c67-hk9db\" (UID: \"36980ce9-53e4-401a-b093-bd1ba933ae83\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.837671 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9gm2\" (UniqueName: \"kubernetes.io/projected/cda39518-2b25-43e2-8afa-b0ff0ec18ff7-kube-api-access-x9gm2\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-8bw69\" (UID: \"cda39518-2b25-43e2-8afa-b0ff0ec18ff7\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.850059 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfbsw\" (UniqueName: \"kubernetes.io/projected/9772f309-5c98-45d7-a404-b14c2ab199e6-kube-api-access-rfbsw\") pod \"octavia-operator-controller-manager-69f59f9d8-qj54s\" (UID: \"9772f309-5c98-45d7-a404-b14c2ab199e6\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.850639 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txwt8\" (UniqueName: \"kubernetes.io/projected/fc43555d-cedf-469e-af48-b2f3cce766c0-kube-api-access-txwt8\") pod \"neutron-operator-controller-manager-69b956fbf6-czwls\" (UID: \"fc43555d-cedf-469e-af48-b2f3cce766c0\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.872949 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.872950 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px247\" (UniqueName: \"kubernetes.io/projected/f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6-kube-api-access-px247\") pod \"ovn-operator-controller-manager-54d485fd9-c5989\" (UID: \"f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6\") " pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.875479 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.877295 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4b7d\" (UniqueName: \"kubernetes.io/projected/0c7c4f04-05a9-40a5-9cb1-9320e33444b1-kube-api-access-w4b7d\") pod \"manila-operator-controller-manager-7cb48dbc-rtfnp\" (UID: \"0c7c4f04-05a9-40a5-9cb1-9320e33444b1\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.891064 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x2c2\" (UniqueName: \"kubernetes.io/projected/e548d82a-0d88-4597-a73e-5e6230e58081-kube-api-access-9x2c2\") pod \"swift-operator-controller-manager-76d5577b-nz284\" (UID: \"e548d82a-0d88-4597-a73e-5e6230e58081\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.891168 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znr2q\" (UniqueName: \"kubernetes.io/projected/aab600a0-d647-44e5-95a4-870d422385ca-kube-api-access-znr2q\") pod \"test-operator-controller-manager-6bb6dcddc-n5vms\" (UID: \"aab600a0-d647-44e5-95a4-870d422385ca\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.891223 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ssjc\" (UniqueName: \"kubernetes.io/projected/0521e121-6b1f-44e3-9823-73f5cb04b5dd-kube-api-access-7ssjc\") pod \"placement-operator-controller-manager-66f6d6849b-ht58x\" (UID: \"0521e121-6b1f-44e3-9823-73f5cb04b5dd\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.891249 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm4v9\" (UniqueName: \"kubernetes.io/projected/6a176a41-220a-4472-9ca5-4e341f86ce49-kube-api-access-mm4v9\") pod \"telemetry-operator-controller-manager-f589c7597-hvgqm\" (UID: \"6a176a41-220a-4472-9ca5-4e341f86ce49\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.914852 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.916383 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.921023 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-6xw67" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.941654 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d"] Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.952362 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.977972 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.992522 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znr2q\" (UniqueName: \"kubernetes.io/projected/aab600a0-d647-44e5-95a4-870d422385ca-kube-api-access-znr2q\") pod \"test-operator-controller-manager-6bb6dcddc-n5vms\" (UID: \"aab600a0-d647-44e5-95a4-870d422385ca\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.992594 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ssjc\" (UniqueName: \"kubernetes.io/projected/0521e121-6b1f-44e3-9823-73f5cb04b5dd-kube-api-access-7ssjc\") pod \"placement-operator-controller-manager-66f6d6849b-ht58x\" (UID: \"0521e121-6b1f-44e3-9823-73f5cb04b5dd\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.992629 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm4v9\" (UniqueName: \"kubernetes.io/projected/6a176a41-220a-4472-9ca5-4e341f86ce49-kube-api-access-mm4v9\") pod \"telemetry-operator-controller-manager-f589c7597-hvgqm\" (UID: \"6a176a41-220a-4472-9ca5-4e341f86ce49\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.992662 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x2c2\" (UniqueName: \"kubernetes.io/projected/e548d82a-0d88-4597-a73e-5e6230e58081-kube-api-access-9x2c2\") pod \"swift-operator-controller-manager-76d5577b-nz284\" (UID: \"e548d82a-0d88-4597-a73e-5e6230e58081\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" Oct 07 22:26:33 crc kubenswrapper[4871]: I1007 22:26:33.994515 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.013985 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x2c2\" (UniqueName: \"kubernetes.io/projected/e548d82a-0d88-4597-a73e-5e6230e58081-kube-api-access-9x2c2\") pod \"swift-operator-controller-manager-76d5577b-nz284\" (UID: \"e548d82a-0d88-4597-a73e-5e6230e58081\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.014473 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.015417 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm4v9\" (UniqueName: \"kubernetes.io/projected/6a176a41-220a-4472-9ca5-4e341f86ce49-kube-api-access-mm4v9\") pod \"telemetry-operator-controller-manager-f589c7597-hvgqm\" (UID: \"6a176a41-220a-4472-9ca5-4e341f86ce49\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.021060 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ssjc\" (UniqueName: \"kubernetes.io/projected/0521e121-6b1f-44e3-9823-73f5cb04b5dd-kube-api-access-7ssjc\") pod \"placement-operator-controller-manager-66f6d6849b-ht58x\" (UID: \"0521e121-6b1f-44e3-9823-73f5cb04b5dd\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.033940 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.034932 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znr2q\" (UniqueName: \"kubernetes.io/projected/aab600a0-d647-44e5-95a4-870d422385ca-kube-api-access-znr2q\") pod \"test-operator-controller-manager-6bb6dcddc-n5vms\" (UID: \"aab600a0-d647-44e5-95a4-870d422385ca\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.040654 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.068230 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs"] Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.071414 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.076140 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.076849 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.077137 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ktvmt" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.082675 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.095477 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs"] Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.096824 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f9kn\" (UniqueName: \"kubernetes.io/projected/ad701ba2-f685-4f23-a42a-dc2480b2e92b-kube-api-access-5f9kn\") pod \"watcher-operator-controller-manager-5d98cc5575-h5h7d\" (UID: \"ad701ba2-f685-4f23-a42a-dc2480b2e92b\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.096920 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-cert\") pod \"infra-operator-controller-manager-84788b6bc5-c6bxk\" (UID: \"e08fd5ec-5b08-4ad9-b298-886bda95d3d9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:34 crc kubenswrapper[4871]: E1007 22:26:34.097129 4871 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 07 22:26:34 crc kubenswrapper[4871]: E1007 22:26:34.097212 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-cert podName:e08fd5ec-5b08-4ad9-b298-886bda95d3d9 nodeName:}" failed. No retries permitted until 2025-10-07 22:26:35.097188015 +0000 UTC m=+1068.899886088 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-cert") pod "infra-operator-controller-manager-84788b6bc5-c6bxk" (UID: "e08fd5ec-5b08-4ad9-b298-886bda95d3d9") : secret "infra-operator-webhook-server-cert" not found Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.110855 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx"] Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.112140 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.124441 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx"] Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.127385 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-cdqn4" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.129021 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.182766 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.199520 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f9kn\" (UniqueName: \"kubernetes.io/projected/ad701ba2-f685-4f23-a42a-dc2480b2e92b-kube-api-access-5f9kn\") pod \"watcher-operator-controller-manager-5d98cc5575-h5h7d\" (UID: \"ad701ba2-f685-4f23-a42a-dc2480b2e92b\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.199597 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15bcef66-2998-4b89-b4ad-18e06aaf57b6-cert\") pod \"openstack-operator-controller-manager-6bfd56c677-s7qzs\" (UID: \"15bcef66-2998-4b89-b4ad-18e06aaf57b6\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.199633 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw2jp\" (UniqueName: \"kubernetes.io/projected/15bcef66-2998-4b89-b4ad-18e06aaf57b6-kube-api-access-cw2jp\") pod \"openstack-operator-controller-manager-6bfd56c677-s7qzs\" (UID: \"15bcef66-2998-4b89-b4ad-18e06aaf57b6\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.259658 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f9kn\" (UniqueName: \"kubernetes.io/projected/ad701ba2-f685-4f23-a42a-dc2480b2e92b-kube-api-access-5f9kn\") pod \"watcher-operator-controller-manager-5d98cc5575-h5h7d\" (UID: \"ad701ba2-f685-4f23-a42a-dc2480b2e92b\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.307127 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f919f707-59de-4f00-a9cf-177ed5696694-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686wj76g\" (UID: \"f919f707-59de-4f00-a9cf-177ed5696694\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.307679 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8g2v\" (UniqueName: \"kubernetes.io/projected/efa19466-3280-43e2-b4a1-2d2077114241-kube-api-access-k8g2v\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx\" (UID: \"efa19466-3280-43e2-b4a1-2d2077114241\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.307716 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15bcef66-2998-4b89-b4ad-18e06aaf57b6-cert\") pod \"openstack-operator-controller-manager-6bfd56c677-s7qzs\" (UID: \"15bcef66-2998-4b89-b4ad-18e06aaf57b6\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:34 crc kubenswrapper[4871]: E1007 22:26:34.307728 4871 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.307746 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw2jp\" (UniqueName: \"kubernetes.io/projected/15bcef66-2998-4b89-b4ad-18e06aaf57b6-kube-api-access-cw2jp\") pod \"openstack-operator-controller-manager-6bfd56c677-s7qzs\" (UID: \"15bcef66-2998-4b89-b4ad-18e06aaf57b6\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:34 crc kubenswrapper[4871]: E1007 22:26:34.307864 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f919f707-59de-4f00-a9cf-177ed5696694-cert podName:f919f707-59de-4f00-a9cf-177ed5696694 nodeName:}" failed. No retries permitted until 2025-10-07 22:26:35.307776976 +0000 UTC m=+1069.110475049 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f919f707-59de-4f00-a9cf-177ed5696694-cert") pod "openstack-baremetal-operator-controller-manager-6875c66686wj76g" (UID: "f919f707-59de-4f00-a9cf-177ed5696694") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 22:26:34 crc kubenswrapper[4871]: E1007 22:26:34.308006 4871 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 07 22:26:34 crc kubenswrapper[4871]: E1007 22:26:34.308086 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15bcef66-2998-4b89-b4ad-18e06aaf57b6-cert podName:15bcef66-2998-4b89-b4ad-18e06aaf57b6 nodeName:}" failed. No retries permitted until 2025-10-07 22:26:34.808065293 +0000 UTC m=+1068.610763366 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/15bcef66-2998-4b89-b4ad-18e06aaf57b6-cert") pod "openstack-operator-controller-manager-6bfd56c677-s7qzs" (UID: "15bcef66-2998-4b89-b4ad-18e06aaf57b6") : secret "webhook-server-cert" not found Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.365218 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw2jp\" (UniqueName: \"kubernetes.io/projected/15bcef66-2998-4b89-b4ad-18e06aaf57b6-kube-api-access-cw2jp\") pod \"openstack-operator-controller-manager-6bfd56c677-s7qzs\" (UID: \"15bcef66-2998-4b89-b4ad-18e06aaf57b6\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.410552 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8g2v\" (UniqueName: \"kubernetes.io/projected/efa19466-3280-43e2-b4a1-2d2077114241-kube-api-access-k8g2v\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx\" (UID: \"efa19466-3280-43e2-b4a1-2d2077114241\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.447635 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8g2v\" (UniqueName: \"kubernetes.io/projected/efa19466-3280-43e2-b4a1-2d2077114241-kube-api-access-k8g2v\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx\" (UID: \"efa19466-3280-43e2-b4a1-2d2077114241\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.537720 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz"] Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.549466 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.550240 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn"] Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.685561 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.821298 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15bcef66-2998-4b89-b4ad-18e06aaf57b6-cert\") pod \"openstack-operator-controller-manager-6bfd56c677-s7qzs\" (UID: \"15bcef66-2998-4b89-b4ad-18e06aaf57b6\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.846555 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15bcef66-2998-4b89-b4ad-18e06aaf57b6-cert\") pod \"openstack-operator-controller-manager-6bfd56c677-s7qzs\" (UID: \"15bcef66-2998-4b89-b4ad-18e06aaf57b6\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.850710 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk"] Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.873476 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql"] Oct 07 22:26:34 crc kubenswrapper[4871]: W1007 22:26:34.917086 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6916e675_a275_4e0f_b43d_c40a8973f13f.slice/crio-edace88ea241da96066eca15798b29613bfdab4366929cf4adea3aa6000e39fb WatchSource:0}: Error finding container edace88ea241da96066eca15798b29613bfdab4366929cf4adea3aa6000e39fb: Status 404 returned error can't find the container with id edace88ea241da96066eca15798b29613bfdab4366929cf4adea3aa6000e39fb Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.931239 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.972728 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97"] Oct 07 22:26:34 crc kubenswrapper[4871]: I1007 22:26:34.999159 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb"] Oct 07 22:26:35 crc kubenswrapper[4871]: W1007 22:26:35.000992 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb58f786d_7c8a_4506_94d4_bf1fcf543871.slice/crio-1ca41170abc44dadb95ef9076b6d7a0a2e50bc1eb7bc15ee98c883f6cd97a946 WatchSource:0}: Error finding container 1ca41170abc44dadb95ef9076b6d7a0a2e50bc1eb7bc15ee98c883f6cd97a946: Status 404 returned error can't find the container with id 1ca41170abc44dadb95ef9076b6d7a0a2e50bc1eb7bc15ee98c883f6cd97a946 Oct 07 22:26:35 crc kubenswrapper[4871]: W1007 22:26:35.010212 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3a4e444_6508_4ae3_8b40_3f238f252ed2.slice/crio-a11ccc547b1246c63b165d52402efc4f7847c8aeeee5f78adaedee717724e2ff WatchSource:0}: Error finding container a11ccc547b1246c63b165d52402efc4f7847c8aeeee5f78adaedee717724e2ff: Status 404 returned error can't find the container with id a11ccc547b1246c63b165d52402efc4f7847c8aeeee5f78adaedee717724e2ff Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.010319 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" event={"ID":"8acf9fc2-6307-4d04-921c-cdfc851f8c52","Type":"ContainerStarted","Data":"b5722f86c6d602a00adc7725a092d2a0b7d5896b5eec2c586510cfc6eda0b56d"} Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.015656 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" event={"ID":"e65e1c1c-1bbc-488e-821c-790bb85bcd55","Type":"ContainerStarted","Data":"099e4f05fbc7b78aa580a07e46e21cfadf577f2f79de55289667ab29ea32517b"} Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.017348 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" event={"ID":"07b6d737-7f0c-494d-adfe-0a641261c344","Type":"ContainerStarted","Data":"a96a5bb1a0eaaf9d3f91398580a99bd8e9b5de63785717d859c9b75131e6ec9f"} Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.018590 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" event={"ID":"6916e675-a275-4e0f-b43d-c40a8973f13f","Type":"ContainerStarted","Data":"edace88ea241da96066eca15798b29613bfdab4366929cf4adea3aa6000e39fb"} Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.126600 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-cert\") pod \"infra-operator-controller-manager-84788b6bc5-c6bxk\" (UID: \"e08fd5ec-5b08-4ad9-b298-886bda95d3d9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.140271 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e08fd5ec-5b08-4ad9-b298-886bda95d3d9-cert\") pod \"infra-operator-controller-manager-84788b6bc5-c6bxk\" (UID: \"e08fd5ec-5b08-4ad9-b298-886bda95d3d9\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.289099 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.311531 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.320257 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:35 crc kubenswrapper[4871]: W1007 22:26:35.323541 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcda39518_2b25_43e2_8afa_b0ff0ec18ff7.slice/crio-2c8c5734a8f0d054988389b97db7e8b602796955c916824564c77485e7df71c7 WatchSource:0}: Error finding container 2c8c5734a8f0d054988389b97db7e8b602796955c916824564c77485e7df71c7: Status 404 returned error can't find the container with id 2c8c5734a8f0d054988389b97db7e8b602796955c916824564c77485e7df71c7 Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.330032 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls"] Oct 07 22:26:35 crc kubenswrapper[4871]: W1007 22:26:35.360464 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b31dae_e86d_4735_96b8_125a9a066a3e.slice/crio-d1afad876b8566846864db428b0ad65f5cf38594bc9b6e6bdf3e1a23afe86488 WatchSource:0}: Error finding container d1afad876b8566846864db428b0ad65f5cf38594bc9b6e6bdf3e1a23afe86488: Status 404 returned error can't find the container with id d1afad876b8566846864db428b0ad65f5cf38594bc9b6e6bdf3e1a23afe86488 Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.367468 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.369602 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f919f707-59de-4f00-a9cf-177ed5696694-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686wj76g\" (UID: \"f919f707-59de-4f00-a9cf-177ed5696694\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.375730 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f919f707-59de-4f00-a9cf-177ed5696694-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686wj76g\" (UID: \"f919f707-59de-4f00-a9cf-177ed5696694\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.381740 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.426803 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.434232 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.437451 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.445586 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.449172 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.452783 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-nz284"] Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.470845 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7ssjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-66f6d6849b-ht58x_openstack-operators(0521e121-6b1f-44e3-9823-73f5cb04b5dd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.471631 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5f9kn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5d98cc5575-h5h7d_openstack-operators(ad701ba2-f685-4f23-a42a-dc2480b2e92b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.481660 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989"] Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.486885 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9x2c2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-76d5577b-nz284_openstack-operators(e548d82a-0d88-4597-a73e-5e6230e58081): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.488069 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d"] Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.488149 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k8g2v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx_openstack-operators(efa19466-3280-43e2-b4a1-2d2077114241): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.491889 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" podUID="efa19466-3280-43e2-b4a1-2d2077114241" Oct 07 22:26:35 crc kubenswrapper[4871]: W1007 22:26:35.492019 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaab600a0_d647_44e5_95a4_870d422385ca.slice/crio-788ca9fa12187a3a0de016467c673f0ffa6adc8f7aa3d00eafd3ac63c3a86afd WatchSource:0}: Error finding container 788ca9fa12187a3a0de016467c673f0ffa6adc8f7aa3d00eafd3ac63c3a86afd: Status 404 returned error can't find the container with id 788ca9fa12187a3a0de016467c673f0ffa6adc8f7aa3d00eafd3ac63c3a86afd Oct 07 22:26:35 crc kubenswrapper[4871]: W1007 22:26:35.492515 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9772f309_5c98_45d7_a404_b14c2ab199e6.slice/crio-7160588d5898197268aeccf526c994fb44cb3e0e60d2cab1a62f964f0cbaa542 WatchSource:0}: Error finding container 7160588d5898197268aeccf526c994fb44cb3e0e60d2cab1a62f964f0cbaa542: Status 404 returned error can't find the container with id 7160588d5898197268aeccf526c994fb44cb3e0e60d2cab1a62f964f0cbaa542 Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.497673 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms"] Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.504317 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.504832 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s"] Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.510306 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-znr2q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-6bb6dcddc-n5vms_openstack-operators(aab600a0-d647-44e5-95a4-870d422385ca): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.512152 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.512202 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.513028 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rfbsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-69f59f9d8-qj54s_openstack-operators(9772f309-5c98-45d7-a404-b14c2ab199e6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.801599 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" podUID="ad701ba2-f685-4f23-a42a-dc2480b2e92b" Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.808588 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" podUID="9772f309-5c98-45d7-a404-b14c2ab199e6" Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.809559 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" podUID="0521e121-6b1f-44e3-9823-73f5cb04b5dd" Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.838229 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" podUID="aab600a0-d647-44e5-95a4-870d422385ca" Oct 07 22:26:35 crc kubenswrapper[4871]: E1007 22:26:35.859336 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" podUID="e548d82a-0d88-4597-a73e-5e6230e58081" Oct 07 22:26:35 crc kubenswrapper[4871]: I1007 22:26:35.933145 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk"] Oct 07 22:26:35 crc kubenswrapper[4871]: W1007 22:26:35.941520 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode08fd5ec_5b08_4ad9_b298_886bda95d3d9.slice/crio-17b834b465504a8b07ace8a557c12fce3df4273668f2331d090322e7c4fa7f5e WatchSource:0}: Error finding container 17b834b465504a8b07ace8a557c12fce3df4273668f2331d090322e7c4fa7f5e: Status 404 returned error can't find the container with id 17b834b465504a8b07ace8a557c12fce3df4273668f2331d090322e7c4fa7f5e Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.030209 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" event={"ID":"e548d82a-0d88-4597-a73e-5e6230e58081","Type":"ContainerStarted","Data":"fe91b9d2b2b82c03c5c7285eda1b8184970399f71dd9019911a74281ae4a4ed5"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.030281 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" event={"ID":"e548d82a-0d88-4597-a73e-5e6230e58081","Type":"ContainerStarted","Data":"9cfd90721b22e63193857629246e8dc3463b60274e3ae98a1e7674a99d168fc4"} Oct 07 22:26:36 crc kubenswrapper[4871]: E1007 22:26:36.032613 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" podUID="e548d82a-0d88-4597-a73e-5e6230e58081" Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.036106 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" event={"ID":"cda39518-2b25-43e2-8afa-b0ff0ec18ff7","Type":"ContainerStarted","Data":"2c8c5734a8f0d054988389b97db7e8b602796955c916824564c77485e7df71c7"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.042332 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" event={"ID":"e85a9c1c-2748-4ae7-a4fa-cfb555348feb","Type":"ContainerStarted","Data":"884b8c89c3255b9cb0409934fae89a0c4935e184488fe4232475ec03d7bfeb9e"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.045672 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" event={"ID":"15bcef66-2998-4b89-b4ad-18e06aaf57b6","Type":"ContainerStarted","Data":"ee1e5725e4bbc153bfc8607fe885693edfcee3d56597e0898fd879858a12ec72"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.045697 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" event={"ID":"15bcef66-2998-4b89-b4ad-18e06aaf57b6","Type":"ContainerStarted","Data":"1f6e4efa82620e7e5db3e58fc7ddd0312f7f29d54977345d68f1c15192aec3b9"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.045712 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" event={"ID":"15bcef66-2998-4b89-b4ad-18e06aaf57b6","Type":"ContainerStarted","Data":"32392b87373ecd97f367a3c9344acb7c1c89fcf7d253182230ac57fcda8843f4"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.046580 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.048018 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" event={"ID":"f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6","Type":"ContainerStarted","Data":"a14bc040999de67aa424f1ae4e7a2b40bb6c2018ef162eed4372fdeee626ff8e"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.050389 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" event={"ID":"0521e121-6b1f-44e3-9823-73f5cb04b5dd","Type":"ContainerStarted","Data":"e8b4ef843b526d5f12abb24ed0e99a6a290c4e51dcaf548c298f279e919a4521"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.050412 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" event={"ID":"0521e121-6b1f-44e3-9823-73f5cb04b5dd","Type":"ContainerStarted","Data":"43cf259533c8bfc048e7abac38252b4724944a2884f8a52714ffe4c0c0fe486f"} Oct 07 22:26:36 crc kubenswrapper[4871]: E1007 22:26:36.051648 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" podUID="0521e121-6b1f-44e3-9823-73f5cb04b5dd" Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.055506 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" event={"ID":"aab600a0-d647-44e5-95a4-870d422385ca","Type":"ContainerStarted","Data":"19eea60492192b7ed7be1a481bea17e463b060aeeae432e9047be4722bb57f2f"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.055577 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" event={"ID":"aab600a0-d647-44e5-95a4-870d422385ca","Type":"ContainerStarted","Data":"788ca9fa12187a3a0de016467c673f0ffa6adc8f7aa3d00eafd3ac63c3a86afd"} Oct 07 22:26:36 crc kubenswrapper[4871]: E1007 22:26:36.061911 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" podUID="aab600a0-d647-44e5-95a4-870d422385ca" Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.065694 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" event={"ID":"0c7c4f04-05a9-40a5-9cb1-9320e33444b1","Type":"ContainerStarted","Data":"6c10d6f5a46cce83bf282874ab93f66af58aba6f1f4f8078d191941e665b5404"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.072333 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" event={"ID":"6a176a41-220a-4472-9ca5-4e341f86ce49","Type":"ContainerStarted","Data":"88f0821a57e28862517b205b5b1ca048ac03517bcc7aee9278ef6864ee348e84"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.073383 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" event={"ID":"36980ce9-53e4-401a-b093-bd1ba933ae83","Type":"ContainerStarted","Data":"36d2f6ce6cd1923df84fe21a8e8ce99f84faa65c4b029da127abce400466985d"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.074210 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" event={"ID":"fc43555d-cedf-469e-af48-b2f3cce766c0","Type":"ContainerStarted","Data":"63deff737df6a9d0011453daee145a8212e4a4874e751e97d00d98392de37580"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.082438 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" event={"ID":"63b31dae-e86d-4735-96b8-125a9a066a3e","Type":"ContainerStarted","Data":"d1afad876b8566846864db428b0ad65f5cf38594bc9b6e6bdf3e1a23afe86488"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.083742 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" event={"ID":"efa19466-3280-43e2-b4a1-2d2077114241","Type":"ContainerStarted","Data":"936702c93ff5240f55c955bac4654ca3e25139ce6980b12ea7a03ec8d6189e87"} Oct 07 22:26:36 crc kubenswrapper[4871]: E1007 22:26:36.087110 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" podUID="efa19466-3280-43e2-b4a1-2d2077114241" Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.087289 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" event={"ID":"b58f786d-7c8a-4506-94d4-bf1fcf543871","Type":"ContainerStarted","Data":"1ca41170abc44dadb95ef9076b6d7a0a2e50bc1eb7bc15ee98c883f6cd97a946"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.097247 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" event={"ID":"e08fd5ec-5b08-4ad9-b298-886bda95d3d9","Type":"ContainerStarted","Data":"17b834b465504a8b07ace8a557c12fce3df4273668f2331d090322e7c4fa7f5e"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.117229 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" event={"ID":"9772f309-5c98-45d7-a404-b14c2ab199e6","Type":"ContainerStarted","Data":"1e7b31d9942913b599414fdfe83f815b8d5fdde1f8044dda2f717545d607a352"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.117268 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" event={"ID":"9772f309-5c98-45d7-a404-b14c2ab199e6","Type":"ContainerStarted","Data":"7160588d5898197268aeccf526c994fb44cb3e0e60d2cab1a62f964f0cbaa542"} Oct 07 22:26:36 crc kubenswrapper[4871]: E1007 22:26:36.119623 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" podUID="9772f309-5c98-45d7-a404-b14c2ab199e6" Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.130483 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" podStartSLOduration=3.1304631179999998 podStartE2EDuration="3.130463118s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:26:36.125454946 +0000 UTC m=+1069.928153019" watchObservedRunningTime="2025-10-07 22:26:36.130463118 +0000 UTC m=+1069.933161191" Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.142935 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" event={"ID":"ad701ba2-f685-4f23-a42a-dc2480b2e92b","Type":"ContainerStarted","Data":"4106b588bdc810eeaa35d27603dd36411c3477e4fa96f9547761345d61f2cfbf"} Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.142984 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" event={"ID":"ad701ba2-f685-4f23-a42a-dc2480b2e92b","Type":"ContainerStarted","Data":"c56696361fb98d57a9785ed51ce0ea94f6d73ced754807d24959c1983486e71a"} Oct 07 22:26:36 crc kubenswrapper[4871]: E1007 22:26:36.153930 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" podUID="ad701ba2-f685-4f23-a42a-dc2480b2e92b" Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.160395 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g"] Oct 07 22:26:36 crc kubenswrapper[4871]: I1007 22:26:36.169503 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" event={"ID":"b3a4e444-6508-4ae3-8b40-3f238f252ed2","Type":"ContainerStarted","Data":"a11ccc547b1246c63b165d52402efc4f7847c8aeeee5f78adaedee717724e2ff"} Oct 07 22:26:37 crc kubenswrapper[4871]: I1007 22:26:37.187167 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" event={"ID":"f919f707-59de-4f00-a9cf-177ed5696694","Type":"ContainerStarted","Data":"9abf30f0c475363c7e0d92de1ae9aee8281c3bedec6ecce0893db91464ce7bf0"} Oct 07 22:26:37 crc kubenswrapper[4871]: E1007 22:26:37.193720 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" podUID="efa19466-3280-43e2-b4a1-2d2077114241" Oct 07 22:26:37 crc kubenswrapper[4871]: E1007 22:26:37.193971 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" podUID="ad701ba2-f685-4f23-a42a-dc2480b2e92b" Oct 07 22:26:37 crc kubenswrapper[4871]: E1007 22:26:37.194050 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" podUID="0521e121-6b1f-44e3-9823-73f5cb04b5dd" Oct 07 22:26:37 crc kubenswrapper[4871]: E1007 22:26:37.194047 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" podUID="9772f309-5c98-45d7-a404-b14c2ab199e6" Oct 07 22:26:37 crc kubenswrapper[4871]: E1007 22:26:37.194062 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" podUID="aab600a0-d647-44e5-95a4-870d422385ca" Oct 07 22:26:37 crc kubenswrapper[4871]: E1007 22:26:37.208083 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" podUID="e548d82a-0d88-4597-a73e-5e6230e58081" Oct 07 22:26:44 crc kubenswrapper[4871]: I1007 22:26:44.940443 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-s7qzs" Oct 07 22:26:47 crc kubenswrapper[4871]: I1007 22:26:47.983493 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.297142 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" event={"ID":"b3a4e444-6508-4ae3-8b40-3f238f252ed2","Type":"ContainerStarted","Data":"a7321811b3183c1069ab0589a567b47e59c4237381b71c687e6b1235c17b343e"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.299063 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" event={"ID":"fc43555d-cedf-469e-af48-b2f3cce766c0","Type":"ContainerStarted","Data":"e0c7b3837b4d824126c0da678eee1105bc069623daed3e25ef964de503938a33"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.305538 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" event={"ID":"07b6d737-7f0c-494d-adfe-0a641261c344","Type":"ContainerStarted","Data":"9e28885eab941ac91a6313768dbd4c529efde0698eb39de9145ecdc0911448a0"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.333135 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" event={"ID":"6916e675-a275-4e0f-b43d-c40a8973f13f","Type":"ContainerStarted","Data":"04d15bfe3cbf69f5d9614b7339e6ce459040e62a7a6638f0ded8da5f30a38329"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.345931 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" event={"ID":"e65e1c1c-1bbc-488e-821c-790bb85bcd55","Type":"ContainerStarted","Data":"b98f7cdb88e18b66dc8da3810271124c7fe80e35fb52e9105ff8afcf73743a65"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.367069 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" event={"ID":"f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6","Type":"ContainerStarted","Data":"9db2384ddef94da219e488d9c7fa8125aea85f334b39541d4b8a88bd56236b44"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.389187 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" event={"ID":"6a176a41-220a-4472-9ca5-4e341f86ce49","Type":"ContainerStarted","Data":"66f99f90cb1f9bf4f0aca774b4335f32b96a4ed6aef7bc0052f46b80bac73eb4"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.411459 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" event={"ID":"e85a9c1c-2748-4ae7-a4fa-cfb555348feb","Type":"ContainerStarted","Data":"296d5ce89fd0319a096fcaa8a9479cf8ac725ce2d37851a2777f8da55098a3ac"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.423648 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" event={"ID":"36980ce9-53e4-401a-b093-bd1ba933ae83","Type":"ContainerStarted","Data":"28d880530ee695a9437ce78358625b22abf90ea57cb8eb9adf84575bc0512951"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.435595 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" event={"ID":"0c7c4f04-05a9-40a5-9cb1-9320e33444b1","Type":"ContainerStarted","Data":"b8e8c0a239e774b9d0425a24f7ba934030a0e18529f68d171b8ed9dfce0f8588"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.445568 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" event={"ID":"b58f786d-7c8a-4506-94d4-bf1fcf543871","Type":"ContainerStarted","Data":"ec9ae0ef2dd6f7a7d14b80bfa1e7746c3304b513fc05a0ee5a1730fc3838f8bf"} Oct 07 22:26:48 crc kubenswrapper[4871]: I1007 22:26:48.449181 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" event={"ID":"e08fd5ec-5b08-4ad9-b298-886bda95d3d9","Type":"ContainerStarted","Data":"3576efbfceac6146b4a4d4009840f519f184dc5ec5ab5f00628c580d52a27ae2"} Oct 07 22:26:49 crc kubenswrapper[4871]: I1007 22:26:49.460903 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" event={"ID":"8acf9fc2-6307-4d04-921c-cdfc851f8c52","Type":"ContainerStarted","Data":"7007ba904a8bf50813bb72322e0d696ec395824ce788a4b508462376bebf13bb"} Oct 07 22:26:49 crc kubenswrapper[4871]: I1007 22:26:49.464978 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" event={"ID":"b3a4e444-6508-4ae3-8b40-3f238f252ed2","Type":"ContainerStarted","Data":"a8921952e80f76591cd9f86ccc23ad63380be12d684d79e294f50e8ba9fb62fc"} Oct 07 22:26:49 crc kubenswrapper[4871]: I1007 22:26:49.465078 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" Oct 07 22:26:49 crc kubenswrapper[4871]: I1007 22:26:49.466919 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" event={"ID":"63b31dae-e86d-4735-96b8-125a9a066a3e","Type":"ContainerStarted","Data":"146b46816d986f85b8a37cb856522ab0b50c9f6c6b4bcacb8a4b6ca0794a44bd"} Oct 07 22:26:49 crc kubenswrapper[4871]: I1007 22:26:49.469019 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" event={"ID":"cda39518-2b25-43e2-8afa-b0ff0ec18ff7","Type":"ContainerStarted","Data":"c9cac13f4ac530cfb49b5f0a21228a2e311ba9906b2badc27240c331a90ce5d5"} Oct 07 22:26:49 crc kubenswrapper[4871]: I1007 22:26:49.486590 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" event={"ID":"f919f707-59de-4f00-a9cf-177ed5696694","Type":"ContainerStarted","Data":"ee31fc63699f429516f0e8c532805dbe8e6db16dd77f4b7add8785bec8f04538"} Oct 07 22:26:49 crc kubenswrapper[4871]: I1007 22:26:49.510200 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" podStartSLOduration=3.9777890830000002 podStartE2EDuration="16.510176679s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.012704551 +0000 UTC m=+1068.815402614" lastFinishedPulling="2025-10-07 22:26:47.545092107 +0000 UTC m=+1081.347790210" observedRunningTime="2025-10-07 22:26:49.506484472 +0000 UTC m=+1083.309182555" watchObservedRunningTime="2025-10-07 22:26:49.510176679 +0000 UTC m=+1083.312874752" Oct 07 22:26:50 crc kubenswrapper[4871]: I1007 22:26:50.501497 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" event={"ID":"fc43555d-cedf-469e-af48-b2f3cce766c0","Type":"ContainerStarted","Data":"26e4f3410a13d6511e20f189e5496448a966e0d9b15ccd0a2e8f2837c5e7806c"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.527319 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" event={"ID":"b58f786d-7c8a-4506-94d4-bf1fcf543871","Type":"ContainerStarted","Data":"a9b4c56374fad5decf912d856f816231b2d6a21beee33e21d58a5f356a3f7fcf"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.529246 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.534254 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.538239 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" event={"ID":"cda39518-2b25-43e2-8afa-b0ff0ec18ff7","Type":"ContainerStarted","Data":"a44e3dfeb3b1f7230110866d5bde5a66ac31096bb0ead6a68f3be4ecf1b9465f"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.538440 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.541447 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" event={"ID":"07b6d737-7f0c-494d-adfe-0a641261c344","Type":"ContainerStarted","Data":"21279ae3aa4db8b2347de38b20554d058c8e97a878ff03b8178fcbf4c9c295f9"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.542126 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.545497 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" event={"ID":"e65e1c1c-1bbc-488e-821c-790bb85bcd55","Type":"ContainerStarted","Data":"e269b5c8d7f13c5cfe34ce71fb51a891e553a8bdebf37a1467c8071588e51ccb"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.545618 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.548191 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-v6x97" podStartSLOduration=6.996794818 podStartE2EDuration="19.548176885s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.003713154 +0000 UTC m=+1068.806411227" lastFinishedPulling="2025-10-07 22:26:47.555095191 +0000 UTC m=+1081.357793294" observedRunningTime="2025-10-07 22:26:52.543764628 +0000 UTC m=+1086.346462701" watchObservedRunningTime="2025-10-07 22:26:52.548176885 +0000 UTC m=+1086.350874958" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.550316 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" event={"ID":"f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6","Type":"ContainerStarted","Data":"a2a74c8d2949c7bd3b2ab8e0bb05a3e6538747bbbcbc7982707ff4fd6ecc749c"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.550635 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.553747 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" event={"ID":"f919f707-59de-4f00-a9cf-177ed5696694","Type":"ContainerStarted","Data":"94c375b63cdeeb8aa60030558cfad6f8d82f389692cebbb371b6364f78181ce8"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.554636 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.570471 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" event={"ID":"e08fd5ec-5b08-4ad9-b298-886bda95d3d9","Type":"ContainerStarted","Data":"12844a47fe745056ffc897d3814a722935c9dc4673cbdc744ead854049a5a533"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.570550 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.574482 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" event={"ID":"8acf9fc2-6307-4d04-921c-cdfc851f8c52","Type":"ContainerStarted","Data":"01af48c6af05ddadd031be8e42184cfc9b10ccc03f389defce90398243bad65a"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.574606 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.577061 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" podStartSLOduration=7.319889939 podStartE2EDuration="19.577043067s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.332667671 +0000 UTC m=+1069.135365744" lastFinishedPulling="2025-10-07 22:26:47.589820799 +0000 UTC m=+1081.392518872" observedRunningTime="2025-10-07 22:26:52.573828682 +0000 UTC m=+1086.376526755" watchObservedRunningTime="2025-10-07 22:26:52.577043067 +0000 UTC m=+1086.379741140" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.588642 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" event={"ID":"6a176a41-220a-4472-9ca5-4e341f86ce49","Type":"ContainerStarted","Data":"79dea772d830024f909b37a701175e6e55858568d39905fa2d8e936c8c01eeca"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.588985 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.605153 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" event={"ID":"e85a9c1c-2748-4ae7-a4fa-cfb555348feb","Type":"ContainerStarted","Data":"6d2aa0dbc006311d4afe858b19627c596a2252f2e688e2326f9b29b00c61dc04"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.605638 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" podStartSLOduration=6.700514373 podStartE2EDuration="19.605615262s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:34.644034655 +0000 UTC m=+1068.446732728" lastFinishedPulling="2025-10-07 22:26:47.549135534 +0000 UTC m=+1081.351833617" observedRunningTime="2025-10-07 22:26:52.601184905 +0000 UTC m=+1086.403882978" watchObservedRunningTime="2025-10-07 22:26:52.605615262 +0000 UTC m=+1086.408313335" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.605734 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.629942 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" event={"ID":"6916e675-a275-4e0f-b43d-c40a8973f13f","Type":"ContainerStarted","Data":"75565ff7858401124a312924fa0d77501267cf5788afbd36cd83c52a5e0d7267"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.630737 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.670570 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" podStartSLOduration=7.501125335 podStartE2EDuration="19.670541276s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.420280044 +0000 UTC m=+1069.222978117" lastFinishedPulling="2025-10-07 22:26:47.589695975 +0000 UTC m=+1081.392394058" observedRunningTime="2025-10-07 22:26:52.662815062 +0000 UTC m=+1086.465513135" watchObservedRunningTime="2025-10-07 22:26:52.670541276 +0000 UTC m=+1086.473239349" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.693447 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" event={"ID":"36980ce9-53e4-401a-b093-bd1ba933ae83","Type":"ContainerStarted","Data":"dad361ca629d782f878f76f55c3bcf7173f33386d893b97a10108acc0dd390f2"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.695068 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.720936 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" event={"ID":"63b31dae-e86d-4735-96b8-125a9a066a3e","Type":"ContainerStarted","Data":"c85f23792b31b2dede9d888aac9bbf501cb93a6610313070f5697c592e7a93f3"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.722486 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" podStartSLOduration=7.575077708 podStartE2EDuration="19.722471568s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.396702521 +0000 UTC m=+1069.199400594" lastFinishedPulling="2025-10-07 22:26:47.544096371 +0000 UTC m=+1081.346794454" observedRunningTime="2025-10-07 22:26:52.684914936 +0000 UTC m=+1086.487613009" watchObservedRunningTime="2025-10-07 22:26:52.722471568 +0000 UTC m=+1086.525169641" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.723339 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.726594 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" podStartSLOduration=8.313763485 podStartE2EDuration="19.726573206s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:36.181560848 +0000 UTC m=+1069.984258921" lastFinishedPulling="2025-10-07 22:26:47.594370569 +0000 UTC m=+1081.397068642" observedRunningTime="2025-10-07 22:26:52.722709594 +0000 UTC m=+1086.525407677" watchObservedRunningTime="2025-10-07 22:26:52.726573206 +0000 UTC m=+1086.529271279" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.728065 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" event={"ID":"0c7c4f04-05a9-40a5-9cb1-9320e33444b1","Type":"ContainerStarted","Data":"fd1e89b74e927f6f94f0d0e7f008f7bc35a0408967ad5e89a539579e6a39da1a"} Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.728112 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.728240 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.796545 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" podStartSLOduration=7.142480864 podStartE2EDuration="19.796524583s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:34.935365769 +0000 UTC m=+1068.738063842" lastFinishedPulling="2025-10-07 22:26:47.589409488 +0000 UTC m=+1081.392107561" observedRunningTime="2025-10-07 22:26:52.790139975 +0000 UTC m=+1086.592838058" watchObservedRunningTime="2025-10-07 22:26:52.796524583 +0000 UTC m=+1086.599222656" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.800347 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" podStartSLOduration=7.177117429 podStartE2EDuration="19.800338674s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:34.932146894 +0000 UTC m=+1068.734844967" lastFinishedPulling="2025-10-07 22:26:47.555368139 +0000 UTC m=+1081.358066212" observedRunningTime="2025-10-07 22:26:52.762661079 +0000 UTC m=+1086.565359152" watchObservedRunningTime="2025-10-07 22:26:52.800338674 +0000 UTC m=+1086.603036747" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.817523 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" podStartSLOduration=6.907253452 podStartE2EDuration="19.817510637s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:34.644688773 +0000 UTC m=+1068.447386846" lastFinishedPulling="2025-10-07 22:26:47.554945948 +0000 UTC m=+1081.357644031" observedRunningTime="2025-10-07 22:26:52.816993994 +0000 UTC m=+1086.619692067" watchObservedRunningTime="2025-10-07 22:26:52.817510637 +0000 UTC m=+1086.620208710" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.838320 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" podStartSLOduration=8.197857575 podStartE2EDuration="19.838305087s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.948979426 +0000 UTC m=+1069.751677499" lastFinishedPulling="2025-10-07 22:26:47.589426928 +0000 UTC m=+1081.392125011" observedRunningTime="2025-10-07 22:26:52.834122756 +0000 UTC m=+1086.636820829" watchObservedRunningTime="2025-10-07 22:26:52.838305087 +0000 UTC m=+1086.641003160" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.885180 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" podStartSLOduration=7.766294418 podStartE2EDuration="19.885162394s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.466027272 +0000 UTC m=+1069.268725345" lastFinishedPulling="2025-10-07 22:26:47.584895248 +0000 UTC m=+1081.387593321" observedRunningTime="2025-10-07 22:26:52.865993818 +0000 UTC m=+1086.668691891" watchObservedRunningTime="2025-10-07 22:26:52.885162394 +0000 UTC m=+1086.687860467" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.888757 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" podStartSLOduration=7.649179365 podStartE2EDuration="19.888749959s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.366387301 +0000 UTC m=+1069.169085374" lastFinishedPulling="2025-10-07 22:26:47.605957905 +0000 UTC m=+1081.408655968" observedRunningTime="2025-10-07 22:26:52.884082915 +0000 UTC m=+1086.686780988" watchObservedRunningTime="2025-10-07 22:26:52.888749959 +0000 UTC m=+1086.691448032" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.906069 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" podStartSLOduration=7.811068531 podStartE2EDuration="19.906052286s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.466012402 +0000 UTC m=+1069.268710475" lastFinishedPulling="2025-10-07 22:26:47.560996157 +0000 UTC m=+1081.363694230" observedRunningTime="2025-10-07 22:26:52.904662799 +0000 UTC m=+1086.707360872" watchObservedRunningTime="2025-10-07 22:26:52.906052286 +0000 UTC m=+1086.708750359" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.927210 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" podStartSLOduration=7.72166999 podStartE2EDuration="19.927184514s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.366384921 +0000 UTC m=+1069.169082994" lastFinishedPulling="2025-10-07 22:26:47.571899445 +0000 UTC m=+1081.374597518" observedRunningTime="2025-10-07 22:26:52.925231032 +0000 UTC m=+1086.727929145" watchObservedRunningTime="2025-10-07 22:26:52.927184514 +0000 UTC m=+1086.729882587" Oct 07 22:26:52 crc kubenswrapper[4871]: I1007 22:26:52.953240 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" podStartSLOduration=7.710196197 podStartE2EDuration="19.953207931s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.341419742 +0000 UTC m=+1069.144117815" lastFinishedPulling="2025-10-07 22:26:47.584431466 +0000 UTC m=+1081.387129549" observedRunningTime="2025-10-07 22:26:52.944675925 +0000 UTC m=+1086.747374009" watchObservedRunningTime="2025-10-07 22:26:52.953207931 +0000 UTC m=+1086.755906014" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.526234 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-fh5vz" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.554586 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-5ndrn" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.642766 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-lg6tk" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.695033 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-jd9ql" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.738069 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-8bw69" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.738398 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rtfnp" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.738576 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-ftlvg" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.739431 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-hk9db" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.740221 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-fd648f65-fsrt7" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.741641 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-czwls" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.741883 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686wj76g" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.741942 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-c5989" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.741974 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-c6bxk" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.742357 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-hvgqm" Oct 07 22:26:53 crc kubenswrapper[4871]: I1007 22:26:53.886580 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-fwjnb" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.780908 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" event={"ID":"ad701ba2-f685-4f23-a42a-dc2480b2e92b","Type":"ContainerStarted","Data":"d18412ae094bb712c5701d38075d7f8ce7b1627b1969e4d6293fe0419923085e"} Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.781957 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.783969 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" event={"ID":"aab600a0-d647-44e5-95a4-870d422385ca","Type":"ContainerStarted","Data":"0aed4bd1753601037248a2d37170da1e15b51a382029f92e204e499c2657ca73"} Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.784243 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.786622 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" event={"ID":"efa19466-3280-43e2-b4a1-2d2077114241","Type":"ContainerStarted","Data":"07fcf8752aec4304eb58604f0af8c1208f537a9f963a0f6ab2241783dad66baf"} Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.789315 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" event={"ID":"e548d82a-0d88-4597-a73e-5e6230e58081","Type":"ContainerStarted","Data":"1ee220d26afa29a3b4bf79e88ac80be7663729e1a5e67f4fff849ca73d13b647"} Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.789579 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.793875 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" event={"ID":"0521e121-6b1f-44e3-9823-73f5cb04b5dd","Type":"ContainerStarted","Data":"4ff66fca3d6978223a9bfe3c4233a158cd945a99c1be2641590fcbcfd812bfb1"} Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.794145 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.797370 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" event={"ID":"9772f309-5c98-45d7-a404-b14c2ab199e6","Type":"ContainerStarted","Data":"b33ab13c1217060b9c3a803a3a2236f711d41d120ac637c6280775687a0935b0"} Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.797693 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.811950 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" podStartSLOduration=3.067924807 podStartE2EDuration="24.811924116s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.471547028 +0000 UTC m=+1069.274245101" lastFinishedPulling="2025-10-07 22:26:57.215546337 +0000 UTC m=+1091.018244410" observedRunningTime="2025-10-07 22:26:57.808246259 +0000 UTC m=+1091.610944352" watchObservedRunningTime="2025-10-07 22:26:57.811924116 +0000 UTC m=+1091.614622179" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.835058 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" podStartSLOduration=3.088083329 podStartE2EDuration="24.835031147s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.470686265 +0000 UTC m=+1069.273384328" lastFinishedPulling="2025-10-07 22:26:57.217634073 +0000 UTC m=+1091.020332146" observedRunningTime="2025-10-07 22:26:57.827644811 +0000 UTC m=+1091.630342884" watchObservedRunningTime="2025-10-07 22:26:57.835031147 +0000 UTC m=+1091.637729250" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.856668 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx" podStartSLOduration=2.102414368 podStartE2EDuration="23.856641967s" podCreationTimestamp="2025-10-07 22:26:34 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.488047944 +0000 UTC m=+1069.290746007" lastFinishedPulling="2025-10-07 22:26:57.242275533 +0000 UTC m=+1091.044973606" observedRunningTime="2025-10-07 22:26:57.851242665 +0000 UTC m=+1091.653940748" watchObservedRunningTime="2025-10-07 22:26:57.856641967 +0000 UTC m=+1091.659340080" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.888670 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" podStartSLOduration=3.185392679 podStartE2EDuration="24.888647102s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.512855529 +0000 UTC m=+1069.315553602" lastFinishedPulling="2025-10-07 22:26:57.216109952 +0000 UTC m=+1091.018808025" observedRunningTime="2025-10-07 22:26:57.880735883 +0000 UTC m=+1091.683433976" watchObservedRunningTime="2025-10-07 22:26:57.888647102 +0000 UTC m=+1091.691345185" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.951292 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" podStartSLOduration=3.2060138240000002 podStartE2EDuration="24.951266336s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.510055185 +0000 UTC m=+1069.312753258" lastFinishedPulling="2025-10-07 22:26:57.255307697 +0000 UTC m=+1091.058005770" observedRunningTime="2025-10-07 22:26:57.921969562 +0000 UTC m=+1091.724667645" watchObservedRunningTime="2025-10-07 22:26:57.951266336 +0000 UTC m=+1091.753964409" Oct 07 22:26:57 crc kubenswrapper[4871]: I1007 22:26:57.959034 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" podStartSLOduration=3.229251777 podStartE2EDuration="24.95900828s" podCreationTimestamp="2025-10-07 22:26:33 +0000 UTC" firstStartedPulling="2025-10-07 22:26:35.48676194 +0000 UTC m=+1069.289460013" lastFinishedPulling="2025-10-07 22:26:57.216518443 +0000 UTC m=+1091.019216516" observedRunningTime="2025-10-07 22:26:57.940101421 +0000 UTC m=+1091.742799494" watchObservedRunningTime="2025-10-07 22:26:57.95900828 +0000 UTC m=+1091.761706353" Oct 07 22:27:04 crc kubenswrapper[4871]: I1007 22:27:04.038948 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-qj54s" Oct 07 22:27:04 crc kubenswrapper[4871]: I1007 22:27:04.044564 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-ht58x" Oct 07 22:27:04 crc kubenswrapper[4871]: I1007 22:27:04.134694 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-76d5577b-nz284" Oct 07 22:27:04 crc kubenswrapper[4871]: I1007 22:27:04.188242 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-n5vms" Oct 07 22:27:04 crc kubenswrapper[4871]: I1007 22:27:04.555547 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-h5h7d" Oct 07 22:27:05 crc kubenswrapper[4871]: I1007 22:27:05.512297 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:27:05 crc kubenswrapper[4871]: I1007 22:27:05.512395 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.536570 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-kthlz"] Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.548994 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.556683 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-kthlz"] Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.557306 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.557306 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-2l4cc" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.557390 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.557518 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.584013 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzjnw\" (UniqueName: \"kubernetes.io/projected/4ee74997-50ba-4e01-89bd-dc75f9df10c8-kube-api-access-fzjnw\") pod \"dnsmasq-dns-7bfcb9d745-kthlz\" (UID: \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\") " pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.584096 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee74997-50ba-4e01-89bd-dc75f9df10c8-config\") pod \"dnsmasq-dns-7bfcb9d745-kthlz\" (UID: \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\") " pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.668404 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-k2kr4"] Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.669951 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.672363 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.686116 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzjnw\" (UniqueName: \"kubernetes.io/projected/4ee74997-50ba-4e01-89bd-dc75f9df10c8-kube-api-access-fzjnw\") pod \"dnsmasq-dns-7bfcb9d745-kthlz\" (UID: \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\") " pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.686191 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee74997-50ba-4e01-89bd-dc75f9df10c8-config\") pod \"dnsmasq-dns-7bfcb9d745-kthlz\" (UID: \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\") " pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.687523 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee74997-50ba-4e01-89bd-dc75f9df10c8-config\") pod \"dnsmasq-dns-7bfcb9d745-kthlz\" (UID: \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\") " pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.691900 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-k2kr4"] Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.718394 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzjnw\" (UniqueName: \"kubernetes.io/projected/4ee74997-50ba-4e01-89bd-dc75f9df10c8-kube-api-access-fzjnw\") pod \"dnsmasq-dns-7bfcb9d745-kthlz\" (UID: \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\") " pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.787101 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-dns-svc\") pod \"dnsmasq-dns-758b79db4c-k2kr4\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.787147 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-config\") pod \"dnsmasq-dns-758b79db4c-k2kr4\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.787166 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bdjc\" (UniqueName: \"kubernetes.io/projected/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-kube-api-access-2bdjc\") pod \"dnsmasq-dns-758b79db4c-k2kr4\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.877672 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.888681 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-dns-svc\") pod \"dnsmasq-dns-758b79db4c-k2kr4\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.888747 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-config\") pod \"dnsmasq-dns-758b79db4c-k2kr4\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.888768 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bdjc\" (UniqueName: \"kubernetes.io/projected/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-kube-api-access-2bdjc\") pod \"dnsmasq-dns-758b79db4c-k2kr4\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.889528 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-dns-svc\") pod \"dnsmasq-dns-758b79db4c-k2kr4\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.889552 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-config\") pod \"dnsmasq-dns-758b79db4c-k2kr4\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.913114 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bdjc\" (UniqueName: \"kubernetes.io/projected/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-kube-api-access-2bdjc\") pod \"dnsmasq-dns-758b79db4c-k2kr4\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:19 crc kubenswrapper[4871]: I1007 22:27:19.992370 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:20 crc kubenswrapper[4871]: I1007 22:27:20.150989 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-kthlz"] Oct 07 22:27:20 crc kubenswrapper[4871]: I1007 22:27:20.527954 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-k2kr4"] Oct 07 22:27:20 crc kubenswrapper[4871]: W1007 22:27:20.535298 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45f6fb8f_a25c_4afe_aea6_c8a28b1fc101.slice/crio-f69327aa3195a2e049ff0e1709e238f670dae1823b43fa0bca7d92d71ff33438 WatchSource:0}: Error finding container f69327aa3195a2e049ff0e1709e238f670dae1823b43fa0bca7d92d71ff33438: Status 404 returned error can't find the container with id f69327aa3195a2e049ff0e1709e238f670dae1823b43fa0bca7d92d71ff33438 Oct 07 22:27:21 crc kubenswrapper[4871]: I1007 22:27:21.047167 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" event={"ID":"4ee74997-50ba-4e01-89bd-dc75f9df10c8","Type":"ContainerStarted","Data":"06b515d2bcb1609a3e5e6168960c3fd51f24baff54cfb4f50382bea0da1257f1"} Oct 07 22:27:21 crc kubenswrapper[4871]: I1007 22:27:21.049003 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" event={"ID":"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101","Type":"ContainerStarted","Data":"f69327aa3195a2e049ff0e1709e238f670dae1823b43fa0bca7d92d71ff33438"} Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.377828 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-k2kr4"] Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.419441 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-644597f84c-7dpr6"] Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.420904 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.429011 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-7dpr6"] Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.546014 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-dns-svc\") pod \"dnsmasq-dns-644597f84c-7dpr6\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.546091 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-config\") pod \"dnsmasq-dns-644597f84c-7dpr6\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.546177 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s6j4\" (UniqueName: \"kubernetes.io/projected/5a771dbf-efc4-44e7-a949-dff9154a2ae3-kube-api-access-4s6j4\") pod \"dnsmasq-dns-644597f84c-7dpr6\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.647617 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-dns-svc\") pod \"dnsmasq-dns-644597f84c-7dpr6\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.647665 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-config\") pod \"dnsmasq-dns-644597f84c-7dpr6\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.647718 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s6j4\" (UniqueName: \"kubernetes.io/projected/5a771dbf-efc4-44e7-a949-dff9154a2ae3-kube-api-access-4s6j4\") pod \"dnsmasq-dns-644597f84c-7dpr6\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.648609 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-dns-svc\") pod \"dnsmasq-dns-644597f84c-7dpr6\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.648695 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-config\") pod \"dnsmasq-dns-644597f84c-7dpr6\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.675530 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s6j4\" (UniqueName: \"kubernetes.io/projected/5a771dbf-efc4-44e7-a949-dff9154a2ae3-kube-api-access-4s6j4\") pod \"dnsmasq-dns-644597f84c-7dpr6\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.678418 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-kthlz"] Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.747408 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77597f887-5fgkc"] Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.752881 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.766116 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-5fgkc"] Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.766322 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.851857 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-config\") pod \"dnsmasq-dns-77597f887-5fgkc\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.851966 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbwbp\" (UniqueName: \"kubernetes.io/projected/40299418-16ab-40a0-9ee4-9d165d6c6e01-kube-api-access-vbwbp\") pod \"dnsmasq-dns-77597f887-5fgkc\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.852079 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-dns-svc\") pod \"dnsmasq-dns-77597f887-5fgkc\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.953836 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbwbp\" (UniqueName: \"kubernetes.io/projected/40299418-16ab-40a0-9ee4-9d165d6c6e01-kube-api-access-vbwbp\") pod \"dnsmasq-dns-77597f887-5fgkc\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.954317 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-dns-svc\") pod \"dnsmasq-dns-77597f887-5fgkc\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.954357 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-config\") pod \"dnsmasq-dns-77597f887-5fgkc\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.955665 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-config\") pod \"dnsmasq-dns-77597f887-5fgkc\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.955708 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-dns-svc\") pod \"dnsmasq-dns-77597f887-5fgkc\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:22 crc kubenswrapper[4871]: I1007 22:27:22.995972 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbwbp\" (UniqueName: \"kubernetes.io/projected/40299418-16ab-40a0-9ee4-9d165d6c6e01-kube-api-access-vbwbp\") pod \"dnsmasq-dns-77597f887-5fgkc\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.117655 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.312156 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-7dpr6"] Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.545266 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.546769 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.556366 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.556557 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.556965 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.557189 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.558289 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.560119 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zckhs" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.560121 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.591661 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.637717 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-5fgkc"] Oct 07 22:27:23 crc kubenswrapper[4871]: W1007 22:27:23.651530 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40299418_16ab_40a0_9ee4_9d165d6c6e01.slice/crio-9e8d58c8da5016b4c66d37db4a3a85d5d63ede45b91ebaa8301e3282b5a215dd WatchSource:0}: Error finding container 9e8d58c8da5016b4c66d37db4a3a85d5d63ede45b91ebaa8301e3282b5a215dd: Status 404 returned error can't find the container with id 9e8d58c8da5016b4c66d37db4a3a85d5d63ede45b91ebaa8301e3282b5a215dd Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669018 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669065 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669092 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669134 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669160 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669192 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669216 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77kcp\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-kube-api-access-77kcp\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669234 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669257 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669358 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.669436 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772196 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772243 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772269 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772317 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772337 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772362 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772388 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77kcp\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-kube-api-access-77kcp\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772409 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772428 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772450 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772471 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.772857 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.773416 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.774538 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.774907 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.775624 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.777019 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.781850 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.782604 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.791244 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.791996 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.794929 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77kcp\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-kube-api-access-77kcp\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.840885 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.841263 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.848380 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.853827 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.854129 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.854337 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.854394 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.854558 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.855504 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.855665 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-64q87" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.863921 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.874612 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.874686 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.874724 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g27pd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-kube-api-access-g27pd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.874758 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.874787 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.874904 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.874939 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.874976 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.875013 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.875038 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.875074 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.904626 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.979004 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.979726 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.979766 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.979820 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.979851 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.979873 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.979936 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.979989 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.980017 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.980038 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g27pd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-kube-api-access-g27pd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.980061 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.984190 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.984816 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.985367 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.986226 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.986277 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.986535 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.987406 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.988034 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:23 crc kubenswrapper[4871]: I1007 22:27:23.989121 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.000824 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g27pd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-kube-api-access-g27pd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.003986 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.013571 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.091995 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-5fgkc" event={"ID":"40299418-16ab-40a0-9ee4-9d165d6c6e01","Type":"ContainerStarted","Data":"9e8d58c8da5016b4c66d37db4a3a85d5d63ede45b91ebaa8301e3282b5a215dd"} Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.093643 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" event={"ID":"5a771dbf-efc4-44e7-a949-dff9154a2ae3","Type":"ContainerStarted","Data":"f9c4046edbe408b979f3a1a8436356052c69a3255d69f0ed6264db22a9b02ee3"} Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.198590 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.425768 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.683609 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 22:27:24 crc kubenswrapper[4871]: W1007 22:27:24.701180 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a30eed1_ba36_43aa_9c02_3e4ebe4e02cb.slice/crio-8e9d23a5a31ea2bce14ff1e207052c1d0d2645872d3ba0dbe8e8de2d8a8c684e WatchSource:0}: Error finding container 8e9d23a5a31ea2bce14ff1e207052c1d0d2645872d3ba0dbe8e8de2d8a8c684e: Status 404 returned error can't find the container with id 8e9d23a5a31ea2bce14ff1e207052c1d0d2645872d3ba0dbe8e8de2d8a8c684e Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.852228 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.853577 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.860977 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.861223 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-dmlcx" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.861490 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.861558 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.861770 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.868512 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.868875 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.995978 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-default\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.996025 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.996097 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-secrets\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.996138 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.996162 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.996308 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.996413 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.996440 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-kolla-config\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:24 crc kubenswrapper[4871]: I1007 22:27:24.996481 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb5vn\" (UniqueName: \"kubernetes.io/projected/54142d12-b692-4acd-83f7-e3c029b69d3a-kube-api-access-zb5vn\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.101843 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.101921 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.101948 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.101992 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.102010 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-kolla-config\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.113311 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb5vn\" (UniqueName: \"kubernetes.io/projected/54142d12-b692-4acd-83f7-e3c029b69d3a-kube-api-access-zb5vn\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.113417 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-default\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.113448 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.113540 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-secrets\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.115246 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.115307 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.121950 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.122519 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-kolla-config\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.126418 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.126984 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-default\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.127784 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.131343 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-secrets\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.152868 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dd77ac8e-36b6-4e16-869a-3ce6bff223ac","Type":"ContainerStarted","Data":"7ab7035a520d5d9b271e84ffbaa98a26ad959ea341166ea18f363c62cd14e0f4"} Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.157897 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb","Type":"ContainerStarted","Data":"8e9d23a5a31ea2bce14ff1e207052c1d0d2645872d3ba0dbe8e8de2d8a8c684e"} Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.186291 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb5vn\" (UniqueName: \"kubernetes.io/projected/54142d12-b692-4acd-83f7-e3c029b69d3a-kube-api-access-zb5vn\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.186580 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " pod="openstack/openstack-galera-0" Oct 07 22:27:25 crc kubenswrapper[4871]: I1007 22:27:25.491422 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.054708 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 22:27:26 crc kubenswrapper[4871]: W1007 22:27:26.066417 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54142d12_b692_4acd_83f7_e3c029b69d3a.slice/crio-cbbf66cd67f9bbcaa232bbb87909cf67f8090f8e8a6840833b437243a2d72314 WatchSource:0}: Error finding container cbbf66cd67f9bbcaa232bbb87909cf67f8090f8e8a6840833b437243a2d72314: Status 404 returned error can't find the container with id cbbf66cd67f9bbcaa232bbb87909cf67f8090f8e8a6840833b437243a2d72314 Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.167168 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54142d12-b692-4acd-83f7-e3c029b69d3a","Type":"ContainerStarted","Data":"cbbf66cd67f9bbcaa232bbb87909cf67f8090f8e8a6840833b437243a2d72314"} Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.479008 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.485690 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.488673 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-srscw" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.491728 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.492117 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.507641 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.509092 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.529866 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.531395 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.538893 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.539092 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.539165 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-kh2hz" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.554823 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646432 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646494 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646521 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d22dl\" (UniqueName: \"kubernetes.io/projected/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kube-api-access-d22dl\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646550 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646569 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646596 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646609 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-config-data\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646627 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646646 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646662 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646690 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kolla-config\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646708 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646727 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.646744 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxzl4\" (UniqueName: \"kubernetes.io/projected/e1207599-6a0d-4019-83c0-8c0d6bed143b-kube-api-access-xxzl4\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748293 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748366 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748394 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d22dl\" (UniqueName: \"kubernetes.io/projected/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kube-api-access-d22dl\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748426 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748447 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748537 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-config-data\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748557 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748578 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748598 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748693 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748728 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kolla-config\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748752 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748770 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.748810 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxzl4\" (UniqueName: \"kubernetes.io/projected/e1207599-6a0d-4019-83c0-8c0d6bed143b-kube-api-access-xxzl4\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.750288 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.750539 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.751236 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.751769 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kolla-config\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.752682 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.752960 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.753549 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-config-data\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.762269 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.762362 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.763459 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.768176 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.771042 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxzl4\" (UniqueName: \"kubernetes.io/projected/e1207599-6a0d-4019-83c0-8c0d6bed143b-kube-api-access-xxzl4\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.776611 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.778738 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d22dl\" (UniqueName: \"kubernetes.io/projected/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kube-api-access-d22dl\") pod \"memcached-0\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " pod="openstack/memcached-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.784876 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.832691 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:26 crc kubenswrapper[4871]: I1007 22:27:26.864295 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 22:27:28 crc kubenswrapper[4871]: I1007 22:27:28.362773 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:27:28 crc kubenswrapper[4871]: I1007 22:27:28.364267 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 22:27:28 crc kubenswrapper[4871]: I1007 22:27:28.368353 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-55mtx" Oct 07 22:27:28 crc kubenswrapper[4871]: I1007 22:27:28.372040 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:27:28 crc kubenswrapper[4871]: I1007 22:27:28.496508 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9xjj\" (UniqueName: \"kubernetes.io/projected/bf921c4f-b397-4a7b-b93b-b999e9cfaa8e-kube-api-access-d9xjj\") pod \"kube-state-metrics-0\" (UID: \"bf921c4f-b397-4a7b-b93b-b999e9cfaa8e\") " pod="openstack/kube-state-metrics-0" Oct 07 22:27:28 crc kubenswrapper[4871]: I1007 22:27:28.598251 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9xjj\" (UniqueName: \"kubernetes.io/projected/bf921c4f-b397-4a7b-b93b-b999e9cfaa8e-kube-api-access-d9xjj\") pod \"kube-state-metrics-0\" (UID: \"bf921c4f-b397-4a7b-b93b-b999e9cfaa8e\") " pod="openstack/kube-state-metrics-0" Oct 07 22:27:28 crc kubenswrapper[4871]: I1007 22:27:28.619061 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9xjj\" (UniqueName: \"kubernetes.io/projected/bf921c4f-b397-4a7b-b93b-b999e9cfaa8e-kube-api-access-d9xjj\") pod \"kube-state-metrics-0\" (UID: \"bf921c4f-b397-4a7b-b93b-b999e9cfaa8e\") " pod="openstack/kube-state-metrics-0" Oct 07 22:27:28 crc kubenswrapper[4871]: I1007 22:27:28.689663 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.167568 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8vjd9"] Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.170054 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.177744 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.177975 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.178015 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-6q6xv" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.184948 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8vjd9"] Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.219885 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-log-ovn\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.220006 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run-ovn\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.220028 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-ovn-controller-tls-certs\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.220079 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-combined-ca-bundle\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.220122 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.220139 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22210cb6-0be2-4f3d-9e04-d36274391d54-scripts\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.220164 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f22hp\" (UniqueName: \"kubernetes.io/projected/22210cb6-0be2-4f3d-9e04-d36274391d54-kube-api-access-f22hp\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.258956 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-jhh4r"] Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.260884 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.289262 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jhh4r"] Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.321142 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run-ovn\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.321714 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-ovn-controller-tls-certs\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.321991 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-lib\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322075 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fc4p\" (UniqueName: \"kubernetes.io/projected/57b1b2c5-f33c-4af0-9707-168eae931b2d-kube-api-access-2fc4p\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322179 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-combined-ca-bundle\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322270 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322341 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22210cb6-0be2-4f3d-9e04-d36274391d54-scripts\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322419 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f22hp\" (UniqueName: \"kubernetes.io/projected/22210cb6-0be2-4f3d-9e04-d36274391d54-kube-api-access-f22hp\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322659 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-log\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322741 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-run\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322829 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-etc-ovs\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322914 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-log-ovn\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.323024 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57b1b2c5-f33c-4af0-9707-168eae931b2d-scripts\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.322468 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run-ovn\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.323413 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.323570 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-log-ovn\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.325708 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22210cb6-0be2-4f3d-9e04-d36274391d54-scripts\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.328620 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-ovn-controller-tls-certs\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.337957 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-combined-ca-bundle\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.340370 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f22hp\" (UniqueName: \"kubernetes.io/projected/22210cb6-0be2-4f3d-9e04-d36274391d54-kube-api-access-f22hp\") pod \"ovn-controller-8vjd9\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.424376 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-log\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.424749 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-run\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.424771 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-etc-ovs\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.424841 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57b1b2c5-f33c-4af0-9707-168eae931b2d-scripts\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.424878 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-lib\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.424899 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fc4p\" (UniqueName: \"kubernetes.io/projected/57b1b2c5-f33c-4af0-9707-168eae931b2d-kube-api-access-2fc4p\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.424656 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-log\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.425012 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-run\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.425305 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-etc-ovs\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.427235 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57b1b2c5-f33c-4af0-9707-168eae931b2d-scripts\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.427382 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-lib\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.441890 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fc4p\" (UniqueName: \"kubernetes.io/projected/57b1b2c5-f33c-4af0-9707-168eae931b2d-kube-api-access-2fc4p\") pod \"ovn-controller-ovs-jhh4r\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.541497 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:33 crc kubenswrapper[4871]: I1007 22:27:33.580699 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.080423 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.082662 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.086303 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.087640 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.087750 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.087934 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-lwhc8" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.088248 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.090102 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.143903 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.144021 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9692529-979b-407e-9826-d93875feda52-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.144090 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24kmx\" (UniqueName: \"kubernetes.io/projected/f9692529-979b-407e-9826-d93875feda52-kube-api-access-24kmx\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.144138 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.144191 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.144267 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.144312 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-config\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.144379 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.246606 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.246775 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.246840 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9692529-979b-407e-9826-d93875feda52-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.246891 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24kmx\" (UniqueName: \"kubernetes.io/projected/f9692529-979b-407e-9826-d93875feda52-kube-api-access-24kmx\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.246927 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.246975 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.247038 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.247070 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-config\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.247777 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.248064 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9692529-979b-407e-9826-d93875feda52-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.248643 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-config\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.252300 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.254220 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.254228 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.254414 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.267962 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24kmx\" (UniqueName: \"kubernetes.io/projected/f9692529-979b-407e-9826-d93875feda52-kube-api-access-24kmx\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.281101 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:34 crc kubenswrapper[4871]: I1007 22:27:34.421838 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:35 crc kubenswrapper[4871]: I1007 22:27:35.512172 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:27:35 crc kubenswrapper[4871]: I1007 22:27:35.512585 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:27:35 crc kubenswrapper[4871]: I1007 22:27:35.512659 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:27:35 crc kubenswrapper[4871]: I1007 22:27:35.513505 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e0ffdee3d72686448819835e9ee3e3402bc2e2e3b339676f30e5a058e9fac0be"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:27:35 crc kubenswrapper[4871]: I1007 22:27:35.513561 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://e0ffdee3d72686448819835e9ee3e3402bc2e2e3b339676f30e5a058e9fac0be" gracePeriod=600 Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.132219 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.133637 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.141388 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.141711 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.142037 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-qmktc" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.142518 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.159644 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.193170 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.193255 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-config\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.193313 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.193348 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.193507 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.193573 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84wtx\" (UniqueName: \"kubernetes.io/projected/0a535bdf-4211-4f42-8137-ee421d6a7fae-kube-api-access-84wtx\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.193613 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.193763 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.286392 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="e0ffdee3d72686448819835e9ee3e3402bc2e2e3b339676f30e5a058e9fac0be" exitCode=0 Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.286457 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"e0ffdee3d72686448819835e9ee3e3402bc2e2e3b339676f30e5a058e9fac0be"} Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.286540 4871 scope.go:117] "RemoveContainer" containerID="5fbc94adc3f0888dbb1bfd5173d9407ea8c5c758052b4cd3dbe78d9b527973c4" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.296032 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.296165 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.296207 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-config\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.296258 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.296285 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.296314 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.296338 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84wtx\" (UniqueName: \"kubernetes.io/projected/0a535bdf-4211-4f42-8137-ee421d6a7fae-kube-api-access-84wtx\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.296364 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.296426 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.297908 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-config\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.298603 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.299934 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.302836 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.305635 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.306531 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.324595 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84wtx\" (UniqueName: \"kubernetes.io/projected/0a535bdf-4211-4f42-8137-ee421d6a7fae-kube-api-access-84wtx\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.326555 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:36 crc kubenswrapper[4871]: I1007 22:27:36.474949 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.920707 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.921945 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2bdjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-758b79db4c-k2kr4_openstack(45f6fb8f-a25c-4afe-aea6-c8a28b1fc101): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.925108 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" podUID="45f6fb8f-a25c-4afe-aea6-c8a28b1fc101" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.938827 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.939022 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4s6j4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-644597f84c-7dpr6_openstack(5a771dbf-efc4-44e7-a949-dff9154a2ae3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.940920 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" podUID="5a771dbf-efc4-44e7-a949-dff9154a2ae3" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.982561 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.982857 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vbwbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-77597f887-5fgkc_openstack(40299418-16ab-40a0-9ee4-9d165d6c6e01): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.985108 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-77597f887-5fgkc" podUID="40299418-16ab-40a0-9ee4-9d165d6c6e01" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.994919 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.995155 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fzjnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bfcb9d745-kthlz_openstack(4ee74997-50ba-4e01-89bd-dc75f9df10c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 22:27:39 crc kubenswrapper[4871]: E1007 22:27:39.996480 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" podUID="4ee74997-50ba-4e01-89bd-dc75f9df10c8" Oct 07 22:27:40 crc kubenswrapper[4871]: E1007 22:27:40.323979 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-77597f887-5fgkc" podUID="40299418-16ab-40a0-9ee4-9d165d6c6e01" Oct 07 22:27:40 crc kubenswrapper[4871]: E1007 22:27:40.324418 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" podUID="5a771dbf-efc4-44e7-a949-dff9154a2ae3" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.245220 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.345250 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-dns-svc\") pod \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.345350 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bdjc\" (UniqueName: \"kubernetes.io/projected/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-kube-api-access-2bdjc\") pod \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.345440 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-config\") pod \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\" (UID: \"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101\") " Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.346492 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-config" (OuterVolumeSpecName: "config") pod "45f6fb8f-a25c-4afe-aea6-c8a28b1fc101" (UID: "45f6fb8f-a25c-4afe-aea6-c8a28b1fc101"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.346929 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "45f6fb8f-a25c-4afe-aea6-c8a28b1fc101" (UID: "45f6fb8f-a25c-4afe-aea6-c8a28b1fc101"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.356611 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-kube-api-access-2bdjc" (OuterVolumeSpecName: "kube-api-access-2bdjc") pod "45f6fb8f-a25c-4afe-aea6-c8a28b1fc101" (UID: "45f6fb8f-a25c-4afe-aea6-c8a28b1fc101"). InnerVolumeSpecName "kube-api-access-2bdjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.362239 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" event={"ID":"45f6fb8f-a25c-4afe-aea6-c8a28b1fc101","Type":"ContainerDied","Data":"f69327aa3195a2e049ff0e1709e238f670dae1823b43fa0bca7d92d71ff33438"} Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.362333 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-k2kr4" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.378459 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" event={"ID":"4ee74997-50ba-4e01-89bd-dc75f9df10c8","Type":"ContainerDied","Data":"06b515d2bcb1609a3e5e6168960c3fd51f24baff54cfb4f50382bea0da1257f1"} Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.378513 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06b515d2bcb1609a3e5e6168960c3fd51f24baff54cfb4f50382bea0da1257f1" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.379280 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.446386 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzjnw\" (UniqueName: \"kubernetes.io/projected/4ee74997-50ba-4e01-89bd-dc75f9df10c8-kube-api-access-fzjnw\") pod \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\" (UID: \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\") " Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.446435 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee74997-50ba-4e01-89bd-dc75f9df10c8-config\") pod \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\" (UID: \"4ee74997-50ba-4e01-89bd-dc75f9df10c8\") " Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.446634 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bdjc\" (UniqueName: \"kubernetes.io/projected/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-kube-api-access-2bdjc\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.446669 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.446681 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.453370 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ee74997-50ba-4e01-89bd-dc75f9df10c8-config" (OuterVolumeSpecName: "config") pod "4ee74997-50ba-4e01-89bd-dc75f9df10c8" (UID: "4ee74997-50ba-4e01-89bd-dc75f9df10c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.459480 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ee74997-50ba-4e01-89bd-dc75f9df10c8-kube-api-access-fzjnw" (OuterVolumeSpecName: "kube-api-access-fzjnw") pod "4ee74997-50ba-4e01-89bd-dc75f9df10c8" (UID: "4ee74997-50ba-4e01-89bd-dc75f9df10c8"). InnerVolumeSpecName "kube-api-access-fzjnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.489216 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-k2kr4"] Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.498586 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-k2kr4"] Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.548767 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzjnw\" (UniqueName: \"kubernetes.io/projected/4ee74997-50ba-4e01-89bd-dc75f9df10c8-kube-api-access-fzjnw\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.548829 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee74997-50ba-4e01-89bd-dc75f9df10c8-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:43 crc kubenswrapper[4871]: E1007 22:27:43.568014 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45f6fb8f_a25c_4afe_aea6_c8a28b1fc101.slice\": RecentStats: unable to find data in memory cache]" Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.615181 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 07 22:27:43 crc kubenswrapper[4871]: W1007 22:27:43.618959 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf602e068_0d3e_4f32_91ca_3ec36c965a7c.slice/crio-47f00885da8c1cfefb3e371701cc2c79f37aaf083da95cc13e2c2698888acca7 WatchSource:0}: Error finding container 47f00885da8c1cfefb3e371701cc2c79f37aaf083da95cc13e2c2698888acca7: Status 404 returned error can't find the container with id 47f00885da8c1cfefb3e371701cc2c79f37aaf083da95cc13e2c2698888acca7 Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.621309 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 22:27:43 crc kubenswrapper[4871]: W1007 22:27:43.621482 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1207599_6a0d_4019_83c0_8c0d6bed143b.slice/crio-b152470c4739f052ea68c8b5f32ead55450b0adba08d66c2d110efa3590e6304 WatchSource:0}: Error finding container b152470c4739f052ea68c8b5f32ead55450b0adba08d66c2d110efa3590e6304: Status 404 returned error can't find the container with id b152470c4739f052ea68c8b5f32ead55450b0adba08d66c2d110efa3590e6304 Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.757723 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8vjd9"] Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.772065 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:27:43 crc kubenswrapper[4871]: W1007 22:27:43.783377 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22210cb6_0be2_4f3d_9e04_d36274391d54.slice/crio-a46269d9afc65812aeb877c434e200be29562fb25baecd4383a37363622aa094 WatchSource:0}: Error finding container a46269d9afc65812aeb877c434e200be29562fb25baecd4383a37363622aa094: Status 404 returned error can't find the container with id a46269d9afc65812aeb877c434e200be29562fb25baecd4383a37363622aa094 Oct 07 22:27:43 crc kubenswrapper[4871]: I1007 22:27:43.894418 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.388324 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54142d12-b692-4acd-83f7-e3c029b69d3a","Type":"ContainerStarted","Data":"ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232"} Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.389758 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f9692529-979b-407e-9826-d93875feda52","Type":"ContainerStarted","Data":"1f2bc1440a98fedcc969594e5cf11b2d39214fdc0054993e38321b50dd3f4099"} Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.393092 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f602e068-0d3e-4f32-91ca-3ec36c965a7c","Type":"ContainerStarted","Data":"47f00885da8c1cfefb3e371701cc2c79f37aaf083da95cc13e2c2698888acca7"} Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.405287 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e1207599-6a0d-4019-83c0-8c0d6bed143b","Type":"ContainerStarted","Data":"ae69967c6d8b30455e625a813a9b799217766f4ad4774d0faf2565e9178b0ee4"} Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.405376 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e1207599-6a0d-4019-83c0-8c0d6bed143b","Type":"ContainerStarted","Data":"b152470c4739f052ea68c8b5f32ead55450b0adba08d66c2d110efa3590e6304"} Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.407789 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8vjd9" event={"ID":"22210cb6-0be2-4f3d-9e04-d36274391d54","Type":"ContainerStarted","Data":"a46269d9afc65812aeb877c434e200be29562fb25baecd4383a37363622aa094"} Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.412818 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"a21e914393b384baa70031a146416b2ffe44375e4ba7b43b6aad9a92b38797c6"} Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.415741 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-kthlz" Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.416337 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bf921c4f-b397-4a7b-b93b-b999e9cfaa8e","Type":"ContainerStarted","Data":"fdb8ee6e311c27b3ed0db1a20a25bf2eada6c9179e05c16bbae212999cc4ed02"} Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.552060 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-kthlz"] Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.559176 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-kthlz"] Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.797501 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 22:27:44 crc kubenswrapper[4871]: W1007 22:27:44.801664 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a535bdf_4211_4f42_8137_ee421d6a7fae.slice/crio-5320581b609b79547a868c9f4f5e9de107ecc6a2e52c7a7992ce5401278c8638 WatchSource:0}: Error finding container 5320581b609b79547a868c9f4f5e9de107ecc6a2e52c7a7992ce5401278c8638: Status 404 returned error can't find the container with id 5320581b609b79547a868c9f4f5e9de107ecc6a2e52c7a7992ce5401278c8638 Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.894360 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jhh4r"] Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.995411 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45f6fb8f-a25c-4afe-aea6-c8a28b1fc101" path="/var/lib/kubelet/pods/45f6fb8f-a25c-4afe-aea6-c8a28b1fc101/volumes" Oct 07 22:27:44 crc kubenswrapper[4871]: I1007 22:27:44.996040 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ee74997-50ba-4e01-89bd-dc75f9df10c8" path="/var/lib/kubelet/pods/4ee74997-50ba-4e01-89bd-dc75f9df10c8/volumes" Oct 07 22:27:45 crc kubenswrapper[4871]: I1007 22:27:45.432853 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0a535bdf-4211-4f42-8137-ee421d6a7fae","Type":"ContainerStarted","Data":"5320581b609b79547a868c9f4f5e9de107ecc6a2e52c7a7992ce5401278c8638"} Oct 07 22:27:45 crc kubenswrapper[4871]: I1007 22:27:45.436327 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb","Type":"ContainerStarted","Data":"3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907"} Oct 07 22:27:45 crc kubenswrapper[4871]: I1007 22:27:45.438182 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dd77ac8e-36b6-4e16-869a-3ce6bff223ac","Type":"ContainerStarted","Data":"ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2"} Oct 07 22:27:45 crc kubenswrapper[4871]: I1007 22:27:45.440392 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jhh4r" event={"ID":"57b1b2c5-f33c-4af0-9707-168eae931b2d","Type":"ContainerStarted","Data":"71c9fab85a529e5e7054bde6eef98e050dedc47387775d33478e2c1e1b5eafc2"} Oct 07 22:27:47 crc kubenswrapper[4871]: I1007 22:27:47.462117 4871 generic.go:334] "Generic (PLEG): container finished" podID="e1207599-6a0d-4019-83c0-8c0d6bed143b" containerID="ae69967c6d8b30455e625a813a9b799217766f4ad4774d0faf2565e9178b0ee4" exitCode=0 Oct 07 22:27:47 crc kubenswrapper[4871]: I1007 22:27:47.462217 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e1207599-6a0d-4019-83c0-8c0d6bed143b","Type":"ContainerDied","Data":"ae69967c6d8b30455e625a813a9b799217766f4ad4774d0faf2565e9178b0ee4"} Oct 07 22:27:47 crc kubenswrapper[4871]: I1007 22:27:47.465343 4871 generic.go:334] "Generic (PLEG): container finished" podID="54142d12-b692-4acd-83f7-e3c029b69d3a" containerID="ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232" exitCode=0 Oct 07 22:27:47 crc kubenswrapper[4871]: I1007 22:27:47.465404 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54142d12-b692-4acd-83f7-e3c029b69d3a","Type":"ContainerDied","Data":"ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232"} Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.494553 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e1207599-6a0d-4019-83c0-8c0d6bed143b","Type":"ContainerStarted","Data":"91e6f4e8b318f1dbdfe9754e8b4775a9cd6b471b118d254f9da7064f27986523"} Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.498387 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8vjd9" event={"ID":"22210cb6-0be2-4f3d-9e04-d36274391d54","Type":"ContainerStarted","Data":"ae1cc42e1fe96bd37dd99ecb252d123aafccbfb8f26e86ac08399eae8a5f9b5d"} Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.498526 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-8vjd9" Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.502684 4871 generic.go:334] "Generic (PLEG): container finished" podID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerID="50ba48bfa54cc28d807c8ba1423f3e7544d91a400fe134f0353833893c0e892b" exitCode=0 Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.502771 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jhh4r" event={"ID":"57b1b2c5-f33c-4af0-9707-168eae931b2d","Type":"ContainerDied","Data":"50ba48bfa54cc28d807c8ba1423f3e7544d91a400fe134f0353833893c0e892b"} Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.504897 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bf921c4f-b397-4a7b-b93b-b999e9cfaa8e","Type":"ContainerStarted","Data":"9b9a9410b344df51c50c3ea156c2ab98a9b2468099f9730d9e3e914e846bbfa3"} Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.504976 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.506824 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f9692529-979b-407e-9826-d93875feda52","Type":"ContainerStarted","Data":"d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6"} Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.510383 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f602e068-0d3e-4f32-91ca-3ec36c965a7c","Type":"ContainerStarted","Data":"cd6faea6c3b696458cd8465298585cf7714badd7a13ee982186f92932560d733"} Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.511170 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.512340 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0a535bdf-4211-4f42-8137-ee421d6a7fae","Type":"ContainerStarted","Data":"99e861992db1f9c7304f7a584cbddbe67197ed224df105828d0b35bd4e191a37"} Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.514744 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54142d12-b692-4acd-83f7-e3c029b69d3a","Type":"ContainerStarted","Data":"38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e"} Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.530044 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=25.530018662 podStartE2EDuration="25.530018662s" podCreationTimestamp="2025-10-07 22:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:27:50.522405901 +0000 UTC m=+1144.325103974" watchObservedRunningTime="2025-10-07 22:27:50.530018662 +0000 UTC m=+1144.332716735" Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.555022 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=10.53383098 podStartE2EDuration="27.554997221s" podCreationTimestamp="2025-10-07 22:27:23 +0000 UTC" firstStartedPulling="2025-10-07 22:27:26.070419052 +0000 UTC m=+1119.873117125" lastFinishedPulling="2025-10-07 22:27:43.091585293 +0000 UTC m=+1136.894283366" observedRunningTime="2025-10-07 22:27:50.546550678 +0000 UTC m=+1144.349248761" watchObservedRunningTime="2025-10-07 22:27:50.554997221 +0000 UTC m=+1144.357695284" Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.578037 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8vjd9" podStartSLOduration=11.724546925 podStartE2EDuration="17.578012629s" podCreationTimestamp="2025-10-07 22:27:33 +0000 UTC" firstStartedPulling="2025-10-07 22:27:43.786197446 +0000 UTC m=+1137.588895519" lastFinishedPulling="2025-10-07 22:27:49.63966314 +0000 UTC m=+1143.442361223" observedRunningTime="2025-10-07 22:27:50.571872447 +0000 UTC m=+1144.374570520" watchObservedRunningTime="2025-10-07 22:27:50.578012629 +0000 UTC m=+1144.380710712" Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.610067 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=16.705534524 podStartE2EDuration="22.610034635s" podCreationTimestamp="2025-10-07 22:27:28 +0000 UTC" firstStartedPulling="2025-10-07 22:27:43.793709765 +0000 UTC m=+1137.596407838" lastFinishedPulling="2025-10-07 22:27:49.698209866 +0000 UTC m=+1143.500907949" observedRunningTime="2025-10-07 22:27:50.593860098 +0000 UTC m=+1144.396558171" watchObservedRunningTime="2025-10-07 22:27:50.610034635 +0000 UTC m=+1144.412732708" Oct 07 22:27:50 crc kubenswrapper[4871]: I1007 22:27:50.646390 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=19.406626868 podStartE2EDuration="24.646323283s" podCreationTimestamp="2025-10-07 22:27:26 +0000 UTC" firstStartedPulling="2025-10-07 22:27:43.622437792 +0000 UTC m=+1137.425135865" lastFinishedPulling="2025-10-07 22:27:48.862134207 +0000 UTC m=+1142.664832280" observedRunningTime="2025-10-07 22:27:50.64394637 +0000 UTC m=+1144.446644453" watchObservedRunningTime="2025-10-07 22:27:50.646323283 +0000 UTC m=+1144.449021416" Oct 07 22:27:51 crc kubenswrapper[4871]: I1007 22:27:51.528855 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jhh4r" event={"ID":"57b1b2c5-f33c-4af0-9707-168eae931b2d","Type":"ContainerStarted","Data":"e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c"} Oct 07 22:27:51 crc kubenswrapper[4871]: I1007 22:27:51.529334 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jhh4r" event={"ID":"57b1b2c5-f33c-4af0-9707-168eae931b2d","Type":"ContainerStarted","Data":"2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c"} Oct 07 22:27:51 crc kubenswrapper[4871]: I1007 22:27:51.558112 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-jhh4r" podStartSLOduration=13.826158313 podStartE2EDuration="18.55808909s" podCreationTimestamp="2025-10-07 22:27:33 +0000 UTC" firstStartedPulling="2025-10-07 22:27:44.909361996 +0000 UTC m=+1138.712060079" lastFinishedPulling="2025-10-07 22:27:49.641292783 +0000 UTC m=+1143.443990856" observedRunningTime="2025-10-07 22:27:51.55578765 +0000 UTC m=+1145.358485733" watchObservedRunningTime="2025-10-07 22:27:51.55808909 +0000 UTC m=+1145.360787163" Oct 07 22:27:52 crc kubenswrapper[4871]: I1007 22:27:52.536015 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:52 crc kubenswrapper[4871]: I1007 22:27:52.536578 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:27:54 crc kubenswrapper[4871]: I1007 22:27:54.558422 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0a535bdf-4211-4f42-8137-ee421d6a7fae","Type":"ContainerStarted","Data":"d94cfb46e87419008029b8ee502b157dab2ee74b35e86d31ffb0102836fd1648"} Oct 07 22:27:54 crc kubenswrapper[4871]: I1007 22:27:54.560897 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f9692529-979b-407e-9826-d93875feda52","Type":"ContainerStarted","Data":"e287d0f45913ff5f5dd60da81232d17cb6a0dd0bdcea624cee7faad1f4925c41"} Oct 07 22:27:54 crc kubenswrapper[4871]: I1007 22:27:54.597610 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=10.497071256 podStartE2EDuration="19.597576915s" podCreationTimestamp="2025-10-07 22:27:35 +0000 UTC" firstStartedPulling="2025-10-07 22:27:44.806611843 +0000 UTC m=+1138.609309916" lastFinishedPulling="2025-10-07 22:27:53.907117512 +0000 UTC m=+1147.709815575" observedRunningTime="2025-10-07 22:27:54.588928657 +0000 UTC m=+1148.391626770" watchObservedRunningTime="2025-10-07 22:27:54.597576915 +0000 UTC m=+1148.400275018" Oct 07 22:27:54 crc kubenswrapper[4871]: I1007 22:27:54.635549 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.733182303 podStartE2EDuration="21.635531707s" podCreationTimestamp="2025-10-07 22:27:33 +0000 UTC" firstStartedPulling="2025-10-07 22:27:43.984623826 +0000 UTC m=+1137.787321899" lastFinishedPulling="2025-10-07 22:27:53.88697323 +0000 UTC m=+1147.689671303" observedRunningTime="2025-10-07 22:27:54.630136885 +0000 UTC m=+1148.432834968" watchObservedRunningTime="2025-10-07 22:27:54.635531707 +0000 UTC m=+1148.438229780" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.422267 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.489943 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.492629 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.492683 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.568778 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.573502 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-5fgkc" event={"ID":"40299418-16ab-40a0-9ee4-9d165d6c6e01","Type":"ContainerStarted","Data":"25dd27a295a450717f65c052c97ed883b629d4372bd8fdbea51acf856834884b"} Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.579310 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" event={"ID":"5a771dbf-efc4-44e7-a949-dff9154a2ae3","Type":"ContainerStarted","Data":"5ddfd4d1195e7c4d2ab0b1af92949e642a29642ebbd0932cc94f56dad37e8936"} Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.586245 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.659119 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.666033 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.932960 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-7dpr6"] Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.975665 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-vxlsj"] Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.977107 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.981195 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.998056 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-d6lwf"] Oct 07 22:27:55 crc kubenswrapper[4871]: I1007 22:27:55.999543 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.002288 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.002809 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovn-rundir\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.002967 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.003029 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ljmg\" (UniqueName: \"kubernetes.io/projected/ec7b131b-1100-4f58-bf1c-012ab142f03b-kube-api-access-5ljmg\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.003337 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7b131b-1100-4f58-bf1c-012ab142f03b-config\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.003444 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-combined-ca-bundle\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.003573 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovs-rundir\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.006430 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vxlsj"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.020343 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-d6lwf"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106045 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-dns-svc\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106124 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106172 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ljmg\" (UniqueName: \"kubernetes.io/projected/ec7b131b-1100-4f58-bf1c-012ab142f03b-kube-api-access-5ljmg\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106202 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106263 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-config\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106288 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7b131b-1100-4f58-bf1c-012ab142f03b-config\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106317 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-combined-ca-bundle\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106373 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovs-rundir\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106409 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovn-rundir\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.106436 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgp28\" (UniqueName: \"kubernetes.io/projected/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-kube-api-access-cgp28\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.107944 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovs-rundir\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.107985 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovn-rundir\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.108191 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7b131b-1100-4f58-bf1c-012ab142f03b-config\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.113960 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-combined-ca-bundle\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.125362 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.129899 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ljmg\" (UniqueName: \"kubernetes.io/projected/ec7b131b-1100-4f58-bf1c-012ab142f03b-kube-api-access-5ljmg\") pod \"ovn-controller-metrics-vxlsj\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.208458 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgp28\" (UniqueName: \"kubernetes.io/projected/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-kube-api-access-cgp28\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.208538 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-dns-svc\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.208580 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.208646 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-config\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.209725 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.209888 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-dns-svc\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.221246 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-config\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.237633 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgp28\" (UniqueName: \"kubernetes.io/projected/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-kube-api-access-cgp28\") pod \"dnsmasq-dns-6d8fd57975-d6lwf\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.296223 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.298161 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-5fgkc"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.315357 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.328427 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-84d6w"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.330200 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.333587 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.356198 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-84d6w"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.416126 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.416626 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.416710 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86fmp\" (UniqueName: \"kubernetes.io/projected/81485cf9-a1f3-4f21-be2c-a1c186d58e66-kube-api-access-86fmp\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.416771 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-config\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.416902 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.483074 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.518373 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-config\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.518431 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.518464 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.518514 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.518569 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86fmp\" (UniqueName: \"kubernetes.io/projected/81485cf9-a1f3-4f21-be2c-a1c186d58e66-kube-api-access-86fmp\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.520378 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-config\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.520969 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.522494 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.524584 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.543418 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-wqf47"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.545631 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wqf47" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.553051 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86fmp\" (UniqueName: \"kubernetes.io/projected/81485cf9-a1f3-4f21-be2c-a1c186d58e66-kube-api-access-86fmp\") pod \"dnsmasq-dns-bc45f6dcf-84d6w\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.560260 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wqf47"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.593963 4871 generic.go:334] "Generic (PLEG): container finished" podID="40299418-16ab-40a0-9ee4-9d165d6c6e01" containerID="25dd27a295a450717f65c052c97ed883b629d4372bd8fdbea51acf856834884b" exitCode=0 Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.594912 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-5fgkc" event={"ID":"40299418-16ab-40a0-9ee4-9d165d6c6e01","Type":"ContainerDied","Data":"25dd27a295a450717f65c052c97ed883b629d4372bd8fdbea51acf856834884b"} Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.596829 4871 generic.go:334] "Generic (PLEG): container finished" podID="5a771dbf-efc4-44e7-a949-dff9154a2ae3" containerID="5ddfd4d1195e7c4d2ab0b1af92949e642a29642ebbd0932cc94f56dad37e8936" exitCode=0 Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.597016 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" event={"ID":"5a771dbf-efc4-44e7-a949-dff9154a2ae3","Type":"ContainerDied","Data":"5ddfd4d1195e7c4d2ab0b1af92949e642a29642ebbd0932cc94f56dad37e8936"} Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.620707 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkpgc\" (UniqueName: \"kubernetes.io/projected/90be0dd0-a9e9-44c9-8142-4ad0d82646ce-kube-api-access-hkpgc\") pod \"keystone-db-create-wqf47\" (UID: \"90be0dd0-a9e9-44c9-8142-4ad0d82646ce\") " pod="openstack/keystone-db-create-wqf47" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.722912 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkpgc\" (UniqueName: \"kubernetes.io/projected/90be0dd0-a9e9-44c9-8142-4ad0d82646ce-kube-api-access-hkpgc\") pod \"keystone-db-create-wqf47\" (UID: \"90be0dd0-a9e9-44c9-8142-4ad0d82646ce\") " pod="openstack/keystone-db-create-wqf47" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.734953 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.744935 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkpgc\" (UniqueName: \"kubernetes.io/projected/90be0dd0-a9e9-44c9-8142-4ad0d82646ce-kube-api-access-hkpgc\") pod \"keystone-db-create-wqf47\" (UID: \"90be0dd0-a9e9-44c9-8142-4ad0d82646ce\") " pod="openstack/keystone-db-create-wqf47" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.775237 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-7bj77"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.777583 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7bj77" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.788563 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7bj77"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.837577 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.838002 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.842221 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vxlsj"] Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.869010 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.916745 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wqf47" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.922769 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.928013 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qdkx\" (UniqueName: \"kubernetes.io/projected/a4de1cd3-27f7-4a66-b5ec-89e5988b4d01-kube-api-access-8qdkx\") pod \"placement-db-create-7bj77\" (UID: \"a4de1cd3-27f7-4a66-b5ec-89e5988b4d01\") " pod="openstack/placement-db-create-7bj77" Oct 07 22:27:56 crc kubenswrapper[4871]: I1007 22:27:56.987545 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-d6lwf"] Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.062504 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qdkx\" (UniqueName: \"kubernetes.io/projected/a4de1cd3-27f7-4a66-b5ec-89e5988b4d01-kube-api-access-8qdkx\") pod \"placement-db-create-7bj77\" (UID: \"a4de1cd3-27f7-4a66-b5ec-89e5988b4d01\") " pod="openstack/placement-db-create-7bj77" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.085214 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qdkx\" (UniqueName: \"kubernetes.io/projected/a4de1cd3-27f7-4a66-b5ec-89e5988b4d01-kube-api-access-8qdkx\") pod \"placement-db-create-7bj77\" (UID: \"a4de1cd3-27f7-4a66-b5ec-89e5988b4d01\") " pod="openstack/placement-db-create-7bj77" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.095900 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7bj77" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.119271 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.129026 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-94p9x"] Oct 07 22:27:57 crc kubenswrapper[4871]: E1007 22:27:57.129635 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40299418-16ab-40a0-9ee4-9d165d6c6e01" containerName="init" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.129654 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="40299418-16ab-40a0-9ee4-9d165d6c6e01" containerName="init" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.132387 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="40299418-16ab-40a0-9ee4-9d165d6c6e01" containerName="init" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.133252 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-94p9x" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.145470 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-94p9x"] Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.154690 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.268443 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-dns-svc\") pod \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.269003 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-config\") pod \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.269030 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-config\") pod \"40299418-16ab-40a0-9ee4-9d165d6c6e01\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.269085 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s6j4\" (UniqueName: \"kubernetes.io/projected/5a771dbf-efc4-44e7-a949-dff9154a2ae3-kube-api-access-4s6j4\") pod \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\" (UID: \"5a771dbf-efc4-44e7-a949-dff9154a2ae3\") " Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.269127 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-dns-svc\") pod \"40299418-16ab-40a0-9ee4-9d165d6c6e01\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.269230 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbwbp\" (UniqueName: \"kubernetes.io/projected/40299418-16ab-40a0-9ee4-9d165d6c6e01-kube-api-access-vbwbp\") pod \"40299418-16ab-40a0-9ee4-9d165d6c6e01\" (UID: \"40299418-16ab-40a0-9ee4-9d165d6c6e01\") " Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.269565 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zljmk\" (UniqueName: \"kubernetes.io/projected/e124f5ed-d573-43b1-b93a-3d25367ef6b6-kube-api-access-zljmk\") pod \"glance-db-create-94p9x\" (UID: \"e124f5ed-d573-43b1-b93a-3d25367ef6b6\") " pod="openstack/glance-db-create-94p9x" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.276304 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a771dbf-efc4-44e7-a949-dff9154a2ae3-kube-api-access-4s6j4" (OuterVolumeSpecName: "kube-api-access-4s6j4") pod "5a771dbf-efc4-44e7-a949-dff9154a2ae3" (UID: "5a771dbf-efc4-44e7-a949-dff9154a2ae3"). InnerVolumeSpecName "kube-api-access-4s6j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.296051 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40299418-16ab-40a0-9ee4-9d165d6c6e01-kube-api-access-vbwbp" (OuterVolumeSpecName: "kube-api-access-vbwbp") pod "40299418-16ab-40a0-9ee4-9d165d6c6e01" (UID: "40299418-16ab-40a0-9ee4-9d165d6c6e01"). InnerVolumeSpecName "kube-api-access-vbwbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.306484 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-config" (OuterVolumeSpecName: "config") pod "40299418-16ab-40a0-9ee4-9d165d6c6e01" (UID: "40299418-16ab-40a0-9ee4-9d165d6c6e01"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.307114 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5a771dbf-efc4-44e7-a949-dff9154a2ae3" (UID: "5a771dbf-efc4-44e7-a949-dff9154a2ae3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.309246 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-84d6w"] Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.319345 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "40299418-16ab-40a0-9ee4-9d165d6c6e01" (UID: "40299418-16ab-40a0-9ee4-9d165d6c6e01"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.323660 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-config" (OuterVolumeSpecName: "config") pod "5a771dbf-efc4-44e7-a949-dff9154a2ae3" (UID: "5a771dbf-efc4-44e7-a949-dff9154a2ae3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.374540 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zljmk\" (UniqueName: \"kubernetes.io/projected/e124f5ed-d573-43b1-b93a-3d25367ef6b6-kube-api-access-zljmk\") pod \"glance-db-create-94p9x\" (UID: \"e124f5ed-d573-43b1-b93a-3d25367ef6b6\") " pod="openstack/glance-db-create-94p9x" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.374633 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbwbp\" (UniqueName: \"kubernetes.io/projected/40299418-16ab-40a0-9ee4-9d165d6c6e01-kube-api-access-vbwbp\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.374649 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.374661 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a771dbf-efc4-44e7-a949-dff9154a2ae3-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.374670 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.374679 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s6j4\" (UniqueName: \"kubernetes.io/projected/5a771dbf-efc4-44e7-a949-dff9154a2ae3-kube-api-access-4s6j4\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.374687 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40299418-16ab-40a0-9ee4-9d165d6c6e01-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.395220 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zljmk\" (UniqueName: \"kubernetes.io/projected/e124f5ed-d573-43b1-b93a-3d25367ef6b6-kube-api-access-zljmk\") pod \"glance-db-create-94p9x\" (UID: \"e124f5ed-d573-43b1-b93a-3d25367ef6b6\") " pod="openstack/glance-db-create-94p9x" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.475938 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.479449 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-94p9x" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.519200 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wqf47"] Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.547323 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.620402 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vxlsj" event={"ID":"ec7b131b-1100-4f58-bf1c-012ab142f03b","Type":"ContainerStarted","Data":"1dee67987aed975e2f4f7e026b61a444e730059988839ac8f1623131ce02236f"} Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.624088 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-5fgkc" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.624192 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-5fgkc" event={"ID":"40299418-16ab-40a0-9ee4-9d165d6c6e01","Type":"ContainerDied","Data":"9e8d58c8da5016b4c66d37db4a3a85d5d63ede45b91ebaa8301e3282b5a215dd"} Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.624232 4871 scope.go:117] "RemoveContainer" containerID="25dd27a295a450717f65c052c97ed883b629d4372bd8fdbea51acf856834884b" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.642110 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7bj77"] Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.644903 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.644885 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-7dpr6" event={"ID":"5a771dbf-efc4-44e7-a949-dff9154a2ae3","Type":"ContainerDied","Data":"f9c4046edbe408b979f3a1a8436356052c69a3255d69f0ed6264db22a9b02ee3"} Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.656656 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" event={"ID":"81485cf9-a1f3-4f21-be2c-a1c186d58e66","Type":"ContainerStarted","Data":"22f1ee87dd4a5bf218975995c5ec0b64d55b97489d42dc09d75a368b745bfaee"} Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.659269 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wqf47" event={"ID":"90be0dd0-a9e9-44c9-8142-4ad0d82646ce","Type":"ContainerStarted","Data":"c66b6be94c1727d57fb0dd5a7daa996941103d678a930c96393d179538bd6611"} Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.661116 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" event={"ID":"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04","Type":"ContainerStarted","Data":"c82f92c423f0db3bc80a5534f223c96978731598b372753aaa6b00e2a75292c2"} Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.739259 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-7dpr6"] Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.748186 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-7dpr6"] Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.748768 4871 scope.go:117] "RemoveContainer" containerID="5ddfd4d1195e7c4d2ab0b1af92949e642a29642ebbd0932cc94f56dad37e8936" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.751385 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.769605 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-5fgkc"] Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.773778 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.775270 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77597f887-5fgkc"] Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.979018 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 07 22:27:57 crc kubenswrapper[4871]: E1007 22:27:57.979393 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a771dbf-efc4-44e7-a949-dff9154a2ae3" containerName="init" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.979412 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a771dbf-efc4-44e7-a949-dff9154a2ae3" containerName="init" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.979576 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a771dbf-efc4-44e7-a949-dff9154a2ae3" containerName="init" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.980462 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.982281 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-72fcc" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.983302 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.983581 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 07 22:27:57 crc kubenswrapper[4871]: I1007 22:27:57.989660 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.011663 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.055550 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-94p9x"] Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.089755 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.089823 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-config\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.089851 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.089919 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv797\" (UniqueName: \"kubernetes.io/projected/fb7d0be2-8b44-44ed-948d-11e1932d27b4-kube-api-access-wv797\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.089961 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.089993 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.090022 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-scripts\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.192268 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv797\" (UniqueName: \"kubernetes.io/projected/fb7d0be2-8b44-44ed-948d-11e1932d27b4-kube-api-access-wv797\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.192385 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.192455 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.192508 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-scripts\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.192591 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.192631 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-config\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.192665 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.193633 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-scripts\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.194064 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-config\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.194149 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.200196 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.200292 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.200386 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.211594 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv797\" (UniqueName: \"kubernetes.io/projected/fb7d0be2-8b44-44ed-948d-11e1932d27b4-kube-api-access-wv797\") pod \"ovn-northd-0\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.299042 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.675131 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7bj77" event={"ID":"a4de1cd3-27f7-4a66-b5ec-89e5988b4d01","Type":"ContainerStarted","Data":"431538a09b85c97b47c9ebe694f1541d0eeae61bad440eeb0d8de240f0ce0f83"} Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.685302 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-94p9x" event={"ID":"e124f5ed-d573-43b1-b93a-3d25367ef6b6","Type":"ContainerStarted","Data":"9d192bb7f74b91f41f9e3c7b9d40262db70d97b5c2aad7f2aecaeb7d85294dd6"} Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.738238 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.825474 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.870861 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-d6lwf"] Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.894949 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-qg2kn"] Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.900265 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.912974 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-qg2kn"] Oct 07 22:27:58 crc kubenswrapper[4871]: I1007 22:27:58.997301 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40299418-16ab-40a0-9ee4-9d165d6c6e01" path="/var/lib/kubelet/pods/40299418-16ab-40a0-9ee4-9d165d6c6e01/volumes" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.014683 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc7np\" (UniqueName: \"kubernetes.io/projected/82f27d48-b207-4c5d-a981-102a64909f5e-kube-api-access-zc7np\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.014888 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-config\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.014930 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-dns-svc\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.014988 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-sb\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.015052 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-nb\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.035945 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a771dbf-efc4-44e7-a949-dff9154a2ae3" path="/var/lib/kubelet/pods/5a771dbf-efc4-44e7-a949-dff9154a2ae3/volumes" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.116537 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-config\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.116621 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-dns-svc\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.116649 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-sb\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.116693 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-nb\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.116757 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc7np\" (UniqueName: \"kubernetes.io/projected/82f27d48-b207-4c5d-a981-102a64909f5e-kube-api-access-zc7np\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.118262 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-config\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.118837 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-dns-svc\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.119387 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-sb\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.119932 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-nb\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.137155 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc7np\" (UniqueName: \"kubernetes.io/projected/82f27d48-b207-4c5d-a981-102a64909f5e-kube-api-access-zc7np\") pod \"dnsmasq-dns-57f58c7cff-qg2kn\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.244338 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.692809 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fb7d0be2-8b44-44ed-948d-11e1932d27b4","Type":"ContainerStarted","Data":"acbdfc581af6868d99263bc705496804406a76fe949cd0e7fcea2854e1352143"} Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.722256 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-qg2kn"] Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.910245 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.915972 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.919325 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.919537 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.919715 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-vjz5q" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.920354 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 07 22:27:59 crc kubenswrapper[4871]: I1007 22:27:59.940492 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.033369 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.033427 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.033523 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-lock\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.033599 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27nsr\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-kube-api-access-27nsr\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.033637 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-cache\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.135899 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.136244 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: E1007 22:28:00.136485 4871 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 22:28:00 crc kubenswrapper[4871]: E1007 22:28:00.136541 4871 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.136567 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.136746 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-lock\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: E1007 22:28:00.136877 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift podName:bb4b36fd-1835-4beb-81fb-0df36301f700 nodeName:}" failed. No retries permitted until 2025-10-07 22:28:00.636814069 +0000 UTC m=+1154.439512142 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift") pod "swift-storage-0" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700") : configmap "swift-ring-files" not found Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.137288 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27nsr\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-kube-api-access-27nsr\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.137487 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-lock\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.137720 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-cache\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.138449 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-cache\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.165029 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27nsr\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-kube-api-access-27nsr\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.166313 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.193025 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-ktt8n"] Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.194323 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.196348 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.197293 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.204910 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.213364 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-ktt8n"] Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.239881 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb54182-7976-4c3b-b64b-cecaee8a672b-etc-swift\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.239958 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-scripts\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.240019 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-swiftconf\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.240043 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhdkd\" (UniqueName: \"kubernetes.io/projected/fdb54182-7976-4c3b-b64b-cecaee8a672b-kube-api-access-nhdkd\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.240076 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-dispersionconf\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.240104 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-combined-ca-bundle\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.240133 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-ring-data-devices\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.253873 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-ktt8n"] Oct 07 22:28:00 crc kubenswrapper[4871]: E1007 22:28:00.254529 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-nhdkd ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-ktt8n" podUID="fdb54182-7976-4c3b-b64b-cecaee8a672b" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.265492 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-7zpls"] Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.283626 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.311860 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7zpls"] Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342587 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-dispersionconf\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342638 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-etc-swift\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342665 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-swiftconf\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342684 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj9zn\" (UniqueName: \"kubernetes.io/projected/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-kube-api-access-jj9zn\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342708 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-combined-ca-bundle\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342728 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-combined-ca-bundle\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342760 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-ring-data-devices\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342884 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb54182-7976-4c3b-b64b-cecaee8a672b-etc-swift\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342910 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-ring-data-devices\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342958 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-scripts\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.342978 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-scripts\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.343024 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-dispersionconf\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.343050 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-swiftconf\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.343069 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhdkd\" (UniqueName: \"kubernetes.io/projected/fdb54182-7976-4c3b-b64b-cecaee8a672b-kube-api-access-nhdkd\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.344127 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-ring-data-devices\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.344567 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb54182-7976-4c3b-b64b-cecaee8a672b-etc-swift\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.345451 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-scripts\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.349220 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-combined-ca-bundle\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.352587 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-swiftconf\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.357827 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-dispersionconf\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.361635 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhdkd\" (UniqueName: \"kubernetes.io/projected/fdb54182-7976-4c3b-b64b-cecaee8a672b-kube-api-access-nhdkd\") pod \"swift-ring-rebalance-ktt8n\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.444611 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-ring-data-devices\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.444722 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-scripts\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.445230 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-dispersionconf\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.445297 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-etc-swift\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.445323 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-swiftconf\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.445342 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj9zn\" (UniqueName: \"kubernetes.io/projected/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-kube-api-access-jj9zn\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.445371 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-combined-ca-bundle\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.445678 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-ring-data-devices\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.445898 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-scripts\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.446130 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-etc-swift\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.449561 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-dispersionconf\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.451445 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-combined-ca-bundle\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.451717 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-swiftconf\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.471720 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj9zn\" (UniqueName: \"kubernetes.io/projected/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-kube-api-access-jj9zn\") pod \"swift-ring-rebalance-7zpls\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.648137 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:00 crc kubenswrapper[4871]: E1007 22:28:00.648426 4871 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 22:28:00 crc kubenswrapper[4871]: E1007 22:28:00.648692 4871 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 22:28:00 crc kubenswrapper[4871]: E1007 22:28:00.648774 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift podName:bb4b36fd-1835-4beb-81fb-0df36301f700 nodeName:}" failed. No retries permitted until 2025-10-07 22:28:01.648747318 +0000 UTC m=+1155.451445401 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift") pod "swift-storage-0" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700") : configmap "swift-ring-files" not found Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.704076 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" event={"ID":"82f27d48-b207-4c5d-a981-102a64909f5e","Type":"ContainerStarted","Data":"697158c8bd7356ef549520d0934aaddc6ee7119f610b5720f943fc169ffc17d9"} Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.704122 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.717652 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.732028 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.749809 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhdkd\" (UniqueName: \"kubernetes.io/projected/fdb54182-7976-4c3b-b64b-cecaee8a672b-kube-api-access-nhdkd\") pod \"fdb54182-7976-4c3b-b64b-cecaee8a672b\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.749948 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-combined-ca-bundle\") pod \"fdb54182-7976-4c3b-b64b-cecaee8a672b\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.750014 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-dispersionconf\") pod \"fdb54182-7976-4c3b-b64b-cecaee8a672b\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.750075 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-scripts\") pod \"fdb54182-7976-4c3b-b64b-cecaee8a672b\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.750146 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-ring-data-devices\") pod \"fdb54182-7976-4c3b-b64b-cecaee8a672b\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.750188 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb54182-7976-4c3b-b64b-cecaee8a672b-etc-swift\") pod \"fdb54182-7976-4c3b-b64b-cecaee8a672b\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.750258 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-swiftconf\") pod \"fdb54182-7976-4c3b-b64b-cecaee8a672b\" (UID: \"fdb54182-7976-4c3b-b64b-cecaee8a672b\") " Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.750968 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "fdb54182-7976-4c3b-b64b-cecaee8a672b" (UID: "fdb54182-7976-4c3b-b64b-cecaee8a672b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.750979 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-scripts" (OuterVolumeSpecName: "scripts") pod "fdb54182-7976-4c3b-b64b-cecaee8a672b" (UID: "fdb54182-7976-4c3b-b64b-cecaee8a672b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.751118 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdb54182-7976-4c3b-b64b-cecaee8a672b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "fdb54182-7976-4c3b-b64b-cecaee8a672b" (UID: "fdb54182-7976-4c3b-b64b-cecaee8a672b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.756996 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdb54182-7976-4c3b-b64b-cecaee8a672b" (UID: "fdb54182-7976-4c3b-b64b-cecaee8a672b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.757042 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "fdb54182-7976-4c3b-b64b-cecaee8a672b" (UID: "fdb54182-7976-4c3b-b64b-cecaee8a672b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.757559 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb54182-7976-4c3b-b64b-cecaee8a672b-kube-api-access-nhdkd" (OuterVolumeSpecName: "kube-api-access-nhdkd") pod "fdb54182-7976-4c3b-b64b-cecaee8a672b" (UID: "fdb54182-7976-4c3b-b64b-cecaee8a672b"). InnerVolumeSpecName "kube-api-access-nhdkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.757610 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "fdb54182-7976-4c3b-b64b-cecaee8a672b" (UID: "fdb54182-7976-4c3b-b64b-cecaee8a672b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.852637 4871 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.852678 4871 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb54182-7976-4c3b-b64b-cecaee8a672b-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.852693 4871 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.852703 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhdkd\" (UniqueName: \"kubernetes.io/projected/fdb54182-7976-4c3b-b64b-cecaee8a672b-kube-api-access-nhdkd\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.852716 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.852725 4871 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb54182-7976-4c3b-b64b-cecaee8a672b-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:00 crc kubenswrapper[4871]: I1007 22:28:00.852735 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb54182-7976-4c3b-b64b-cecaee8a672b-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.085103 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7zpls"] Oct 07 22:28:01 crc kubenswrapper[4871]: W1007 22:28:01.089751 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb13b0a19_d03a_4a5c_b77d_e95d97c534fd.slice/crio-5309bb1afc9710ffd7a96a9805b6057a599a55af635ebf3ed91aec89082f35b5 WatchSource:0}: Error finding container 5309bb1afc9710ffd7a96a9805b6057a599a55af635ebf3ed91aec89082f35b5: Status 404 returned error can't find the container with id 5309bb1afc9710ffd7a96a9805b6057a599a55af635ebf3ed91aec89082f35b5 Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.674650 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:01 crc kubenswrapper[4871]: E1007 22:28:01.674874 4871 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 22:28:01 crc kubenswrapper[4871]: E1007 22:28:01.674899 4871 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 22:28:01 crc kubenswrapper[4871]: E1007 22:28:01.674946 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift podName:bb4b36fd-1835-4beb-81fb-0df36301f700 nodeName:}" failed. No retries permitted until 2025-10-07 22:28:03.674927687 +0000 UTC m=+1157.477625760 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift") pod "swift-storage-0" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700") : configmap "swift-ring-files" not found Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.715303 4871 generic.go:334] "Generic (PLEG): container finished" podID="4c6d89d3-70dd-4abc-822c-65ca5dbc1c04" containerID="c458aa38e9fabf30355121bb2a0ca6d86888a135448d52d58824028220ecf4bb" exitCode=0 Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.715833 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" event={"ID":"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04","Type":"ContainerDied","Data":"c458aa38e9fabf30355121bb2a0ca6d86888a135448d52d58824028220ecf4bb"} Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.720485 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7bj77" event={"ID":"a4de1cd3-27f7-4a66-b5ec-89e5988b4d01","Type":"ContainerStarted","Data":"ee6089abc123f34af422521b137cc52ea7e71cd2d1835d9222589cfa258c4a2b"} Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.721675 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7zpls" event={"ID":"b13b0a19-d03a-4a5c-b77d-e95d97c534fd","Type":"ContainerStarted","Data":"5309bb1afc9710ffd7a96a9805b6057a599a55af635ebf3ed91aec89082f35b5"} Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.722973 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vxlsj" event={"ID":"ec7b131b-1100-4f58-bf1c-012ab142f03b","Type":"ContainerStarted","Data":"0dce78a091797582beabb0faed2c660e6204b78527df1d3a95b8d0c04c1f1d92"} Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.738984 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" event={"ID":"82f27d48-b207-4c5d-a981-102a64909f5e","Type":"ContainerStarted","Data":"3490a71fe3931fb507ef1f0808f6041cb516b64c97f3f38d2cfa61e0a7b4398f"} Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.757066 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" event={"ID":"81485cf9-a1f3-4f21-be2c-a1c186d58e66","Type":"ContainerStarted","Data":"acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b"} Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.777526 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ktt8n" Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.780196 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wqf47" event={"ID":"90be0dd0-a9e9-44c9-8142-4ad0d82646ce","Type":"ContainerStarted","Data":"24949ae5014b016a913830d4eb1e8127bec0c97db6d4be6265a57b25735e1294"} Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.831038 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-vxlsj" podStartSLOduration=6.831009069 podStartE2EDuration="6.831009069s" podCreationTimestamp="2025-10-07 22:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:01.770345777 +0000 UTC m=+1155.573043850" watchObservedRunningTime="2025-10-07 22:28:01.831009069 +0000 UTC m=+1155.633707142" Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.861641 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-wqf47" podStartSLOduration=5.861618257 podStartE2EDuration="5.861618257s" podCreationTimestamp="2025-10-07 22:27:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:01.85643681 +0000 UTC m=+1155.659134883" watchObservedRunningTime="2025-10-07 22:28:01.861618257 +0000 UTC m=+1155.664316330" Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.955882 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-ktt8n"] Oct 07 22:28:01 crc kubenswrapper[4871]: I1007 22:28:01.962005 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-ktt8n"] Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.478978 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.505342 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-ovsdbserver-nb\") pod \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.505445 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-dns-svc\") pod \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.505654 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-config\") pod \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.505746 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgp28\" (UniqueName: \"kubernetes.io/projected/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-kube-api-access-cgp28\") pod \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\" (UID: \"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04\") " Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.513713 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-kube-api-access-cgp28" (OuterVolumeSpecName: "kube-api-access-cgp28") pod "4c6d89d3-70dd-4abc-822c-65ca5dbc1c04" (UID: "4c6d89d3-70dd-4abc-822c-65ca5dbc1c04"). InnerVolumeSpecName "kube-api-access-cgp28". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.531037 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c6d89d3-70dd-4abc-822c-65ca5dbc1c04" (UID: "4c6d89d3-70dd-4abc-822c-65ca5dbc1c04"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.557435 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-config" (OuterVolumeSpecName: "config") pod "4c6d89d3-70dd-4abc-822c-65ca5dbc1c04" (UID: "4c6d89d3-70dd-4abc-822c-65ca5dbc1c04"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.560488 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4c6d89d3-70dd-4abc-822c-65ca5dbc1c04" (UID: "4c6d89d3-70dd-4abc-822c-65ca5dbc1c04"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.608422 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.608454 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.608468 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.608481 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgp28\" (UniqueName: \"kubernetes.io/projected/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04-kube-api-access-cgp28\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.787047 4871 generic.go:334] "Generic (PLEG): container finished" podID="81485cf9-a1f3-4f21-be2c-a1c186d58e66" containerID="acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b" exitCode=0 Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.787105 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" event={"ID":"81485cf9-a1f3-4f21-be2c-a1c186d58e66","Type":"ContainerDied","Data":"acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b"} Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.789544 4871 generic.go:334] "Generic (PLEG): container finished" podID="90be0dd0-a9e9-44c9-8142-4ad0d82646ce" containerID="24949ae5014b016a913830d4eb1e8127bec0c97db6d4be6265a57b25735e1294" exitCode=0 Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.789586 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wqf47" event={"ID":"90be0dd0-a9e9-44c9-8142-4ad0d82646ce","Type":"ContainerDied","Data":"24949ae5014b016a913830d4eb1e8127bec0c97db6d4be6265a57b25735e1294"} Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.792501 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" event={"ID":"4c6d89d3-70dd-4abc-822c-65ca5dbc1c04","Type":"ContainerDied","Data":"c82f92c423f0db3bc80a5534f223c96978731598b372753aaa6b00e2a75292c2"} Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.792534 4871 scope.go:117] "RemoveContainer" containerID="c458aa38e9fabf30355121bb2a0ca6d86888a135448d52d58824028220ecf4bb" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.792616 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-d6lwf" Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.805075 4871 generic.go:334] "Generic (PLEG): container finished" podID="a4de1cd3-27f7-4a66-b5ec-89e5988b4d01" containerID="ee6089abc123f34af422521b137cc52ea7e71cd2d1835d9222589cfa258c4a2b" exitCode=0 Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.805140 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7bj77" event={"ID":"a4de1cd3-27f7-4a66-b5ec-89e5988b4d01","Type":"ContainerDied","Data":"ee6089abc123f34af422521b137cc52ea7e71cd2d1835d9222589cfa258c4a2b"} Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.815642 4871 generic.go:334] "Generic (PLEG): container finished" podID="e124f5ed-d573-43b1-b93a-3d25367ef6b6" containerID="302e959638c8acd6469fc15e7a7a5fbe4511eec6e971d1b72ea3c8e26f4fea68" exitCode=0 Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.815743 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-94p9x" event={"ID":"e124f5ed-d573-43b1-b93a-3d25367ef6b6","Type":"ContainerDied","Data":"302e959638c8acd6469fc15e7a7a5fbe4511eec6e971d1b72ea3c8e26f4fea68"} Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.821032 4871 generic.go:334] "Generic (PLEG): container finished" podID="82f27d48-b207-4c5d-a981-102a64909f5e" containerID="3490a71fe3931fb507ef1f0808f6041cb516b64c97f3f38d2cfa61e0a7b4398f" exitCode=0 Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.821118 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" event={"ID":"82f27d48-b207-4c5d-a981-102a64909f5e","Type":"ContainerDied","Data":"3490a71fe3931fb507ef1f0808f6041cb516b64c97f3f38d2cfa61e0a7b4398f"} Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.948277 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-d6lwf"] Oct 07 22:28:02 crc kubenswrapper[4871]: I1007 22:28:02.975941 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-d6lwf"] Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.005963 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c6d89d3-70dd-4abc-822c-65ca5dbc1c04" path="/var/lib/kubelet/pods/4c6d89d3-70dd-4abc-822c-65ca5dbc1c04/volumes" Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.006569 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb54182-7976-4c3b-b64b-cecaee8a672b" path="/var/lib/kubelet/pods/fdb54182-7976-4c3b-b64b-cecaee8a672b/volumes" Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.748699 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:03 crc kubenswrapper[4871]: E1007 22:28:03.749718 4871 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 22:28:03 crc kubenswrapper[4871]: E1007 22:28:03.749741 4871 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 22:28:03 crc kubenswrapper[4871]: E1007 22:28:03.749813 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift podName:bb4b36fd-1835-4beb-81fb-0df36301f700 nodeName:}" failed. No retries permitted until 2025-10-07 22:28:07.749780928 +0000 UTC m=+1161.552478991 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift") pod "swift-storage-0" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700") : configmap "swift-ring-files" not found Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.831931 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fb7d0be2-8b44-44ed-948d-11e1932d27b4","Type":"ContainerStarted","Data":"a0bcaf69c69657ecf8f36f03f56417c9e0e10c4b3d9151b2ed50c569b93152aa"} Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.832001 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fb7d0be2-8b44-44ed-948d-11e1932d27b4","Type":"ContainerStarted","Data":"691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303"} Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.832061 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.835010 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" event={"ID":"82f27d48-b207-4c5d-a981-102a64909f5e","Type":"ContainerStarted","Data":"8b399b7d37e73b596db12fe18565964c2f91556cbf97326d9e702af147df24d6"} Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.835101 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.837937 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" event={"ID":"81485cf9-a1f3-4f21-be2c-a1c186d58e66","Type":"ContainerStarted","Data":"e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad"} Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.838111 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.864191 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.864487568 podStartE2EDuration="6.864167308s" podCreationTimestamp="2025-10-07 22:27:57 +0000 UTC" firstStartedPulling="2025-10-07 22:27:58.804394725 +0000 UTC m=+1152.607092788" lastFinishedPulling="2025-10-07 22:28:02.804074455 +0000 UTC m=+1156.606772528" observedRunningTime="2025-10-07 22:28:03.86081637 +0000 UTC m=+1157.663514443" watchObservedRunningTime="2025-10-07 22:28:03.864167308 +0000 UTC m=+1157.666865381" Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.887151 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" podStartSLOduration=7.887122575 podStartE2EDuration="7.887122575s" podCreationTimestamp="2025-10-07 22:27:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:03.884985128 +0000 UTC m=+1157.687683201" watchObservedRunningTime="2025-10-07 22:28:03.887122575 +0000 UTC m=+1157.689820668" Oct 07 22:28:03 crc kubenswrapper[4871]: I1007 22:28:03.916127 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" podStartSLOduration=5.91609672 podStartE2EDuration="5.91609672s" podCreationTimestamp="2025-10-07 22:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:03.908726335 +0000 UTC m=+1157.711424418" watchObservedRunningTime="2025-10-07 22:28:03.91609672 +0000 UTC m=+1157.718794793" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.565094 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7bj77" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.572728 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wqf47" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.593778 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-94p9x" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.668345 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkpgc\" (UniqueName: \"kubernetes.io/projected/90be0dd0-a9e9-44c9-8142-4ad0d82646ce-kube-api-access-hkpgc\") pod \"90be0dd0-a9e9-44c9-8142-4ad0d82646ce\" (UID: \"90be0dd0-a9e9-44c9-8142-4ad0d82646ce\") " Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.668470 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qdkx\" (UniqueName: \"kubernetes.io/projected/a4de1cd3-27f7-4a66-b5ec-89e5988b4d01-kube-api-access-8qdkx\") pod \"a4de1cd3-27f7-4a66-b5ec-89e5988b4d01\" (UID: \"a4de1cd3-27f7-4a66-b5ec-89e5988b4d01\") " Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.675155 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90be0dd0-a9e9-44c9-8142-4ad0d82646ce-kube-api-access-hkpgc" (OuterVolumeSpecName: "kube-api-access-hkpgc") pod "90be0dd0-a9e9-44c9-8142-4ad0d82646ce" (UID: "90be0dd0-a9e9-44c9-8142-4ad0d82646ce"). InnerVolumeSpecName "kube-api-access-hkpgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.677623 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4de1cd3-27f7-4a66-b5ec-89e5988b4d01-kube-api-access-8qdkx" (OuterVolumeSpecName: "kube-api-access-8qdkx") pod "a4de1cd3-27f7-4a66-b5ec-89e5988b4d01" (UID: "a4de1cd3-27f7-4a66-b5ec-89e5988b4d01"). InnerVolumeSpecName "kube-api-access-8qdkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.770231 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zljmk\" (UniqueName: \"kubernetes.io/projected/e124f5ed-d573-43b1-b93a-3d25367ef6b6-kube-api-access-zljmk\") pod \"e124f5ed-d573-43b1-b93a-3d25367ef6b6\" (UID: \"e124f5ed-d573-43b1-b93a-3d25367ef6b6\") " Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.770699 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkpgc\" (UniqueName: \"kubernetes.io/projected/90be0dd0-a9e9-44c9-8142-4ad0d82646ce-kube-api-access-hkpgc\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.770715 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qdkx\" (UniqueName: \"kubernetes.io/projected/a4de1cd3-27f7-4a66-b5ec-89e5988b4d01-kube-api-access-8qdkx\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.777334 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e124f5ed-d573-43b1-b93a-3d25367ef6b6-kube-api-access-zljmk" (OuterVolumeSpecName: "kube-api-access-zljmk") pod "e124f5ed-d573-43b1-b93a-3d25367ef6b6" (UID: "e124f5ed-d573-43b1-b93a-3d25367ef6b6"). InnerVolumeSpecName "kube-api-access-zljmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.851004 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7bj77" event={"ID":"a4de1cd3-27f7-4a66-b5ec-89e5988b4d01","Type":"ContainerDied","Data":"431538a09b85c97b47c9ebe694f1541d0eeae61bad440eeb0d8de240f0ce0f83"} Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.851283 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="431538a09b85c97b47c9ebe694f1541d0eeae61bad440eeb0d8de240f0ce0f83" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.851085 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7bj77" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.865073 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-94p9x" event={"ID":"e124f5ed-d573-43b1-b93a-3d25367ef6b6","Type":"ContainerDied","Data":"9d192bb7f74b91f41f9e3c7b9d40262db70d97b5c2aad7f2aecaeb7d85294dd6"} Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.865155 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d192bb7f74b91f41f9e3c7b9d40262db70d97b5c2aad7f2aecaeb7d85294dd6" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.865556 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-94p9x" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.871444 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wqf47" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.871722 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wqf47" event={"ID":"90be0dd0-a9e9-44c9-8142-4ad0d82646ce","Type":"ContainerDied","Data":"c66b6be94c1727d57fb0dd5a7daa996941103d678a930c96393d179538bd6611"} Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.871951 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c66b6be94c1727d57fb0dd5a7daa996941103d678a930c96393d179538bd6611" Oct 07 22:28:04 crc kubenswrapper[4871]: I1007 22:28:04.872846 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zljmk\" (UniqueName: \"kubernetes.io/projected/e124f5ed-d573-43b1-b93a-3d25367ef6b6-kube-api-access-zljmk\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:06 crc kubenswrapper[4871]: I1007 22:28:06.890447 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7zpls" event={"ID":"b13b0a19-d03a-4a5c-b77d-e95d97c534fd","Type":"ContainerStarted","Data":"dbb6d5909ff3ea7e58ebe374f5b565ded04d14ec719380765bf7aae3bcbbb674"} Oct 07 22:28:06 crc kubenswrapper[4871]: I1007 22:28:06.918470 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-7zpls" podStartSLOduration=2.080574298 podStartE2EDuration="6.918447683s" podCreationTimestamp="2025-10-07 22:28:00 +0000 UTC" firstStartedPulling="2025-10-07 22:28:01.092360323 +0000 UTC m=+1154.895058396" lastFinishedPulling="2025-10-07 22:28:05.930233678 +0000 UTC m=+1159.732931781" observedRunningTime="2025-10-07 22:28:06.913249826 +0000 UTC m=+1160.715947899" watchObservedRunningTime="2025-10-07 22:28:06.918447683 +0000 UTC m=+1160.721145756" Oct 07 22:28:07 crc kubenswrapper[4871]: I1007 22:28:07.831265 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:07 crc kubenswrapper[4871]: E1007 22:28:07.831740 4871 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 07 22:28:07 crc kubenswrapper[4871]: E1007 22:28:07.831832 4871 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 07 22:28:07 crc kubenswrapper[4871]: E1007 22:28:07.831954 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift podName:bb4b36fd-1835-4beb-81fb-0df36301f700 nodeName:}" failed. No retries permitted until 2025-10-07 22:28:15.831912406 +0000 UTC m=+1169.634610519 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift") pod "swift-storage-0" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700") : configmap "swift-ring-files" not found Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.248059 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.327819 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-84d6w"] Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.328047 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" podUID="81485cf9-a1f3-4f21-be2c-a1c186d58e66" containerName="dnsmasq-dns" containerID="cri-o://e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad" gracePeriod=10 Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.333954 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.778356 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.874016 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-sb\") pod \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.874174 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86fmp\" (UniqueName: \"kubernetes.io/projected/81485cf9-a1f3-4f21-be2c-a1c186d58e66-kube-api-access-86fmp\") pod \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.874213 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-nb\") pod \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.874319 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-config\") pod \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.874415 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-dns-svc\") pod \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\" (UID: \"81485cf9-a1f3-4f21-be2c-a1c186d58e66\") " Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.901119 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81485cf9-a1f3-4f21-be2c-a1c186d58e66-kube-api-access-86fmp" (OuterVolumeSpecName: "kube-api-access-86fmp") pod "81485cf9-a1f3-4f21-be2c-a1c186d58e66" (UID: "81485cf9-a1f3-4f21-be2c-a1c186d58e66"). InnerVolumeSpecName "kube-api-access-86fmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.919880 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "81485cf9-a1f3-4f21-be2c-a1c186d58e66" (UID: "81485cf9-a1f3-4f21-be2c-a1c186d58e66"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.921376 4871 generic.go:334] "Generic (PLEG): container finished" podID="81485cf9-a1f3-4f21-be2c-a1c186d58e66" containerID="e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad" exitCode=0 Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.921435 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" event={"ID":"81485cf9-a1f3-4f21-be2c-a1c186d58e66","Type":"ContainerDied","Data":"e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad"} Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.921468 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" event={"ID":"81485cf9-a1f3-4f21-be2c-a1c186d58e66","Type":"ContainerDied","Data":"22f1ee87dd4a5bf218975995c5ec0b64d55b97489d42dc09d75a368b745bfaee"} Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.921488 4871 scope.go:117] "RemoveContainer" containerID="e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.921651 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-84d6w" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.924091 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-config" (OuterVolumeSpecName: "config") pod "81485cf9-a1f3-4f21-be2c-a1c186d58e66" (UID: "81485cf9-a1f3-4f21-be2c-a1c186d58e66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.926940 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "81485cf9-a1f3-4f21-be2c-a1c186d58e66" (UID: "81485cf9-a1f3-4f21-be2c-a1c186d58e66"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.934991 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "81485cf9-a1f3-4f21-be2c-a1c186d58e66" (UID: "81485cf9-a1f3-4f21-be2c-a1c186d58e66"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.965543 4871 scope.go:117] "RemoveContainer" containerID="acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.977533 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.977594 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.977612 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.977626 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86fmp\" (UniqueName: \"kubernetes.io/projected/81485cf9-a1f3-4f21-be2c-a1c186d58e66-kube-api-access-86fmp\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.977639 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81485cf9-a1f3-4f21-be2c-a1c186d58e66-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.996767 4871 scope.go:117] "RemoveContainer" containerID="e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad" Oct 07 22:28:09 crc kubenswrapper[4871]: E1007 22:28:09.997443 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad\": container with ID starting with e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad not found: ID does not exist" containerID="e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.997512 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad"} err="failed to get container status \"e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad\": rpc error: code = NotFound desc = could not find container \"e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad\": container with ID starting with e479f07e9ff45c4bf17c76b28a9bf0c602837326af59218ad6c06c5155c7a6ad not found: ID does not exist" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.997553 4871 scope.go:117] "RemoveContainer" containerID="acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b" Oct 07 22:28:09 crc kubenswrapper[4871]: E1007 22:28:09.998088 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b\": container with ID starting with acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b not found: ID does not exist" containerID="acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b" Oct 07 22:28:09 crc kubenswrapper[4871]: I1007 22:28:09.998125 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b"} err="failed to get container status \"acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b\": rpc error: code = NotFound desc = could not find container \"acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b\": container with ID starting with acf8d293b13fd8770a7d180090d880346a7b4a4f5c528f787c8f3b27d24f611b not found: ID does not exist" Oct 07 22:28:10 crc kubenswrapper[4871]: I1007 22:28:10.260147 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-84d6w"] Oct 07 22:28:10 crc kubenswrapper[4871]: I1007 22:28:10.269740 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-84d6w"] Oct 07 22:28:11 crc kubenswrapper[4871]: I1007 22:28:11.001205 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81485cf9-a1f3-4f21-be2c-a1c186d58e66" path="/var/lib/kubelet/pods/81485cf9-a1f3-4f21-be2c-a1c186d58e66/volumes" Oct 07 22:28:13 crc kubenswrapper[4871]: I1007 22:28:13.398264 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 07 22:28:13 crc kubenswrapper[4871]: I1007 22:28:13.962749 4871 generic.go:334] "Generic (PLEG): container finished" podID="b13b0a19-d03a-4a5c-b77d-e95d97c534fd" containerID="dbb6d5909ff3ea7e58ebe374f5b565ded04d14ec719380765bf7aae3bcbbb674" exitCode=0 Oct 07 22:28:13 crc kubenswrapper[4871]: I1007 22:28:13.962832 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7zpls" event={"ID":"b13b0a19-d03a-4a5c-b77d-e95d97c534fd","Type":"ContainerDied","Data":"dbb6d5909ff3ea7e58ebe374f5b565ded04d14ec719380765bf7aae3bcbbb674"} Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.373102 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.494339 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-etc-swift\") pod \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.494386 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-combined-ca-bundle\") pod \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.494450 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-swiftconf\") pod \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.494473 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-dispersionconf\") pod \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.494527 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-ring-data-devices\") pod \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.494656 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj9zn\" (UniqueName: \"kubernetes.io/projected/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-kube-api-access-jj9zn\") pod \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.494725 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-scripts\") pod \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\" (UID: \"b13b0a19-d03a-4a5c-b77d-e95d97c534fd\") " Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.495763 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b13b0a19-d03a-4a5c-b77d-e95d97c534fd" (UID: "b13b0a19-d03a-4a5c-b77d-e95d97c534fd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.496054 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b13b0a19-d03a-4a5c-b77d-e95d97c534fd" (UID: "b13b0a19-d03a-4a5c-b77d-e95d97c534fd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.526370 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-kube-api-access-jj9zn" (OuterVolumeSpecName: "kube-api-access-jj9zn") pod "b13b0a19-d03a-4a5c-b77d-e95d97c534fd" (UID: "b13b0a19-d03a-4a5c-b77d-e95d97c534fd"). InnerVolumeSpecName "kube-api-access-jj9zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.536087 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b13b0a19-d03a-4a5c-b77d-e95d97c534fd" (UID: "b13b0a19-d03a-4a5c-b77d-e95d97c534fd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.548227 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b13b0a19-d03a-4a5c-b77d-e95d97c534fd" (UID: "b13b0a19-d03a-4a5c-b77d-e95d97c534fd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.557565 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-scripts" (OuterVolumeSpecName: "scripts") pod "b13b0a19-d03a-4a5c-b77d-e95d97c534fd" (UID: "b13b0a19-d03a-4a5c-b77d-e95d97c534fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.558672 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b13b0a19-d03a-4a5c-b77d-e95d97c534fd" (UID: "b13b0a19-d03a-4a5c-b77d-e95d97c534fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.597010 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.597044 4871 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.597057 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.597071 4871 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.597082 4871 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.597091 4871 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.597101 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj9zn\" (UniqueName: \"kubernetes.io/projected/b13b0a19-d03a-4a5c-b77d-e95d97c534fd-kube-api-access-jj9zn\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.903496 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.910485 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") pod \"swift-storage-0\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " pod="openstack/swift-storage-0" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.986367 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7zpls" event={"ID":"b13b0a19-d03a-4a5c-b77d-e95d97c534fd","Type":"ContainerDied","Data":"5309bb1afc9710ffd7a96a9805b6057a599a55af635ebf3ed91aec89082f35b5"} Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.986448 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5309bb1afc9710ffd7a96a9805b6057a599a55af635ebf3ed91aec89082f35b5" Oct 07 22:28:15 crc kubenswrapper[4871]: I1007 22:28:15.986454 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7zpls" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.136720 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.597750 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6eb9-account-create-ngxgg"] Oct 07 22:28:16 crc kubenswrapper[4871]: E1007 22:28:16.599156 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4de1cd3-27f7-4a66-b5ec-89e5988b4d01" containerName="mariadb-database-create" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599179 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4de1cd3-27f7-4a66-b5ec-89e5988b4d01" containerName="mariadb-database-create" Oct 07 22:28:16 crc kubenswrapper[4871]: E1007 22:28:16.599201 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6d89d3-70dd-4abc-822c-65ca5dbc1c04" containerName="init" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599210 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6d89d3-70dd-4abc-822c-65ca5dbc1c04" containerName="init" Oct 07 22:28:16 crc kubenswrapper[4871]: E1007 22:28:16.599223 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e124f5ed-d573-43b1-b93a-3d25367ef6b6" containerName="mariadb-database-create" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599232 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e124f5ed-d573-43b1-b93a-3d25367ef6b6" containerName="mariadb-database-create" Oct 07 22:28:16 crc kubenswrapper[4871]: E1007 22:28:16.599248 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13b0a19-d03a-4a5c-b77d-e95d97c534fd" containerName="swift-ring-rebalance" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599256 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13b0a19-d03a-4a5c-b77d-e95d97c534fd" containerName="swift-ring-rebalance" Oct 07 22:28:16 crc kubenswrapper[4871]: E1007 22:28:16.599267 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90be0dd0-a9e9-44c9-8142-4ad0d82646ce" containerName="mariadb-database-create" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599275 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="90be0dd0-a9e9-44c9-8142-4ad0d82646ce" containerName="mariadb-database-create" Oct 07 22:28:16 crc kubenswrapper[4871]: E1007 22:28:16.599300 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81485cf9-a1f3-4f21-be2c-a1c186d58e66" containerName="init" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599308 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="81485cf9-a1f3-4f21-be2c-a1c186d58e66" containerName="init" Oct 07 22:28:16 crc kubenswrapper[4871]: E1007 22:28:16.599328 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81485cf9-a1f3-4f21-be2c-a1c186d58e66" containerName="dnsmasq-dns" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599336 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="81485cf9-a1f3-4f21-be2c-a1c186d58e66" containerName="dnsmasq-dns" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599541 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="90be0dd0-a9e9-44c9-8142-4ad0d82646ce" containerName="mariadb-database-create" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599593 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4de1cd3-27f7-4a66-b5ec-89e5988b4d01" containerName="mariadb-database-create" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599628 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="81485cf9-a1f3-4f21-be2c-a1c186d58e66" containerName="dnsmasq-dns" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599645 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c6d89d3-70dd-4abc-822c-65ca5dbc1c04" containerName="init" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599673 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e124f5ed-d573-43b1-b93a-3d25367ef6b6" containerName="mariadb-database-create" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.599697 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b13b0a19-d03a-4a5c-b77d-e95d97c534fd" containerName="swift-ring-rebalance" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.600448 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6eb9-account-create-ngxgg" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.604713 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.614009 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6eb9-account-create-ngxgg"] Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.717349 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsgwd\" (UniqueName: \"kubernetes.io/projected/b6594e86-288e-4da2-8057-71b4fa8ed8b3-kube-api-access-hsgwd\") pod \"keystone-6eb9-account-create-ngxgg\" (UID: \"b6594e86-288e-4da2-8057-71b4fa8ed8b3\") " pod="openstack/keystone-6eb9-account-create-ngxgg" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.755328 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 07 22:28:16 crc kubenswrapper[4871]: W1007 22:28:16.762891 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb4b36fd_1835_4beb_81fb_0df36301f700.slice/crio-67cf4c1b020198e8c306f3bf42d00cf8f2a48766602384e5d3238451532dfb48 WatchSource:0}: Error finding container 67cf4c1b020198e8c306f3bf42d00cf8f2a48766602384e5d3238451532dfb48: Status 404 returned error can't find the container with id 67cf4c1b020198e8c306f3bf42d00cf8f2a48766602384e5d3238451532dfb48 Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.819260 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsgwd\" (UniqueName: \"kubernetes.io/projected/b6594e86-288e-4da2-8057-71b4fa8ed8b3-kube-api-access-hsgwd\") pod \"keystone-6eb9-account-create-ngxgg\" (UID: \"b6594e86-288e-4da2-8057-71b4fa8ed8b3\") " pod="openstack/keystone-6eb9-account-create-ngxgg" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.855605 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsgwd\" (UniqueName: \"kubernetes.io/projected/b6594e86-288e-4da2-8057-71b4fa8ed8b3-kube-api-access-hsgwd\") pod \"keystone-6eb9-account-create-ngxgg\" (UID: \"b6594e86-288e-4da2-8057-71b4fa8ed8b3\") " pod="openstack/keystone-6eb9-account-create-ngxgg" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.881483 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f2b9-account-create-p2nlf"] Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.882932 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f2b9-account-create-p2nlf" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.889107 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.903246 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f2b9-account-create-p2nlf"] Oct 07 22:28:16 crc kubenswrapper[4871]: I1007 22:28:16.932727 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6eb9-account-create-ngxgg" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.005723 4871 generic.go:334] "Generic (PLEG): container finished" podID="1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" containerID="3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907" exitCode=0 Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.020081 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"67cf4c1b020198e8c306f3bf42d00cf8f2a48766602384e5d3238451532dfb48"} Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.020124 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb","Type":"ContainerDied","Data":"3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907"} Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.023009 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqgqc\" (UniqueName: \"kubernetes.io/projected/8dfe9a33-d304-430d-90dc-86a532e34080-kube-api-access-lqgqc\") pod \"placement-f2b9-account-create-p2nlf\" (UID: \"8dfe9a33-d304-430d-90dc-86a532e34080\") " pod="openstack/placement-f2b9-account-create-p2nlf" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.125028 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqgqc\" (UniqueName: \"kubernetes.io/projected/8dfe9a33-d304-430d-90dc-86a532e34080-kube-api-access-lqgqc\") pod \"placement-f2b9-account-create-p2nlf\" (UID: \"8dfe9a33-d304-430d-90dc-86a532e34080\") " pod="openstack/placement-f2b9-account-create-p2nlf" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.145512 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqgqc\" (UniqueName: \"kubernetes.io/projected/8dfe9a33-d304-430d-90dc-86a532e34080-kube-api-access-lqgqc\") pod \"placement-f2b9-account-create-p2nlf\" (UID: \"8dfe9a33-d304-430d-90dc-86a532e34080\") " pod="openstack/placement-f2b9-account-create-p2nlf" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.190641 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a79e-account-create-fwv68"] Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.192021 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a79e-account-create-fwv68" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.198714 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.207458 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a79e-account-create-fwv68"] Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.210384 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f2b9-account-create-p2nlf" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.329010 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svm4m\" (UniqueName: \"kubernetes.io/projected/0c59e7ea-43a5-4ae5-bd0d-36bdb761a042-kube-api-access-svm4m\") pod \"glance-a79e-account-create-fwv68\" (UID: \"0c59e7ea-43a5-4ae5-bd0d-36bdb761a042\") " pod="openstack/glance-a79e-account-create-fwv68" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.418613 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6eb9-account-create-ngxgg"] Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.430258 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svm4m\" (UniqueName: \"kubernetes.io/projected/0c59e7ea-43a5-4ae5-bd0d-36bdb761a042-kube-api-access-svm4m\") pod \"glance-a79e-account-create-fwv68\" (UID: \"0c59e7ea-43a5-4ae5-bd0d-36bdb761a042\") " pod="openstack/glance-a79e-account-create-fwv68" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.458252 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svm4m\" (UniqueName: \"kubernetes.io/projected/0c59e7ea-43a5-4ae5-bd0d-36bdb761a042-kube-api-access-svm4m\") pod \"glance-a79e-account-create-fwv68\" (UID: \"0c59e7ea-43a5-4ae5-bd0d-36bdb761a042\") " pod="openstack/glance-a79e-account-create-fwv68" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.510951 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a79e-account-create-fwv68" Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.723453 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f2b9-account-create-p2nlf"] Oct 07 22:28:17 crc kubenswrapper[4871]: I1007 22:28:17.763681 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a79e-account-create-fwv68"] Oct 07 22:28:17 crc kubenswrapper[4871]: W1007 22:28:17.774330 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c59e7ea_43a5_4ae5_bd0d_36bdb761a042.slice/crio-8759537b388dabaf81d62ee6bf88bc248bc1583ac2ae56ab6e7789ae95fbb688 WatchSource:0}: Error finding container 8759537b388dabaf81d62ee6bf88bc248bc1583ac2ae56ab6e7789ae95fbb688: Status 404 returned error can't find the container with id 8759537b388dabaf81d62ee6bf88bc248bc1583ac2ae56ab6e7789ae95fbb688 Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.028443 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a79e-account-create-fwv68" event={"ID":"0c59e7ea-43a5-4ae5-bd0d-36bdb761a042","Type":"ContainerStarted","Data":"3cafaed81f8618ad2d607a8694a1aa368e23d06c9cfb94bab5130c43974b6205"} Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.028508 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a79e-account-create-fwv68" event={"ID":"0c59e7ea-43a5-4ae5-bd0d-36bdb761a042","Type":"ContainerStarted","Data":"8759537b388dabaf81d62ee6bf88bc248bc1583ac2ae56ab6e7789ae95fbb688"} Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.035932 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f2b9-account-create-p2nlf" event={"ID":"8dfe9a33-d304-430d-90dc-86a532e34080","Type":"ContainerStarted","Data":"587e087735b583679a4e0927c1e92977a3a6e911ca84ae72de2b03c4a34930de"} Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.035977 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f2b9-account-create-p2nlf" event={"ID":"8dfe9a33-d304-430d-90dc-86a532e34080","Type":"ContainerStarted","Data":"ecb04e92c3c34fd869e1fc836f2f875fb05cca725a8a4b24142ca050c3f7a1ed"} Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.048171 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb","Type":"ContainerStarted","Data":"798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d"} Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.049266 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.051147 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-a79e-account-create-fwv68" podStartSLOduration=1.051133298 podStartE2EDuration="1.051133298s" podCreationTimestamp="2025-10-07 22:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:18.050529312 +0000 UTC m=+1171.853227385" watchObservedRunningTime="2025-10-07 22:28:18.051133298 +0000 UTC m=+1171.853831381" Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.058364 4871 generic.go:334] "Generic (PLEG): container finished" podID="b6594e86-288e-4da2-8057-71b4fa8ed8b3" containerID="f560fa449772ee68ef8cd0288a01316c4f2453ff2605948fb794728f08420983" exitCode=0 Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.058488 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6eb9-account-create-ngxgg" event={"ID":"b6594e86-288e-4da2-8057-71b4fa8ed8b3","Type":"ContainerDied","Data":"f560fa449772ee68ef8cd0288a01316c4f2453ff2605948fb794728f08420983"} Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.058530 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6eb9-account-create-ngxgg" event={"ID":"b6594e86-288e-4da2-8057-71b4fa8ed8b3","Type":"ContainerStarted","Data":"044ae24734c0f74825c0a3b867bba1e96046c1628dc5ee3d52fc40b28f9807d3"} Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.062338 4871 generic.go:334] "Generic (PLEG): container finished" podID="dd77ac8e-36b6-4e16-869a-3ce6bff223ac" containerID="ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2" exitCode=0 Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.062396 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dd77ac8e-36b6-4e16-869a-3ce6bff223ac","Type":"ContainerDied","Data":"ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2"} Oct 07 22:28:18 crc kubenswrapper[4871]: I1007 22:28:18.107207 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.694427324 podStartE2EDuration="56.107157853s" podCreationTimestamp="2025-10-07 22:27:22 +0000 UTC" firstStartedPulling="2025-10-07 22:27:24.70708232 +0000 UTC m=+1118.509780393" lastFinishedPulling="2025-10-07 22:27:43.119812839 +0000 UTC m=+1136.922510922" observedRunningTime="2025-10-07 22:28:18.098819642 +0000 UTC m=+1171.901517715" watchObservedRunningTime="2025-10-07 22:28:18.107157853 +0000 UTC m=+1171.909856026" Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.091153 4871 generic.go:334] "Generic (PLEG): container finished" podID="0c59e7ea-43a5-4ae5-bd0d-36bdb761a042" containerID="3cafaed81f8618ad2d607a8694a1aa368e23d06c9cfb94bab5130c43974b6205" exitCode=0 Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.091251 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a79e-account-create-fwv68" event={"ID":"0c59e7ea-43a5-4ae5-bd0d-36bdb761a042","Type":"ContainerDied","Data":"3cafaed81f8618ad2d607a8694a1aa368e23d06c9cfb94bab5130c43974b6205"} Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.099907 4871 generic.go:334] "Generic (PLEG): container finished" podID="8dfe9a33-d304-430d-90dc-86a532e34080" containerID="587e087735b583679a4e0927c1e92977a3a6e911ca84ae72de2b03c4a34930de" exitCode=0 Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.100208 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f2b9-account-create-p2nlf" event={"ID":"8dfe9a33-d304-430d-90dc-86a532e34080","Type":"ContainerDied","Data":"587e087735b583679a4e0927c1e92977a3a6e911ca84ae72de2b03c4a34930de"} Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.106189 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dd77ac8e-36b6-4e16-869a-3ce6bff223ac","Type":"ContainerStarted","Data":"fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb"} Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.151406 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.430938984 podStartE2EDuration="57.151386209s" podCreationTimestamp="2025-10-07 22:27:22 +0000 UTC" firstStartedPulling="2025-10-07 22:27:24.448308476 +0000 UTC m=+1118.251006559" lastFinishedPulling="2025-10-07 22:27:43.168755711 +0000 UTC m=+1136.971453784" observedRunningTime="2025-10-07 22:28:19.144734463 +0000 UTC m=+1172.947432576" watchObservedRunningTime="2025-10-07 22:28:19.151386209 +0000 UTC m=+1172.954084292" Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.628853 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f2b9-account-create-p2nlf" Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.636444 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6eb9-account-create-ngxgg" Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.786431 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqgqc\" (UniqueName: \"kubernetes.io/projected/8dfe9a33-d304-430d-90dc-86a532e34080-kube-api-access-lqgqc\") pod \"8dfe9a33-d304-430d-90dc-86a532e34080\" (UID: \"8dfe9a33-d304-430d-90dc-86a532e34080\") " Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.786639 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsgwd\" (UniqueName: \"kubernetes.io/projected/b6594e86-288e-4da2-8057-71b4fa8ed8b3-kube-api-access-hsgwd\") pod \"b6594e86-288e-4da2-8057-71b4fa8ed8b3\" (UID: \"b6594e86-288e-4da2-8057-71b4fa8ed8b3\") " Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.793942 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6594e86-288e-4da2-8057-71b4fa8ed8b3-kube-api-access-hsgwd" (OuterVolumeSpecName: "kube-api-access-hsgwd") pod "b6594e86-288e-4da2-8057-71b4fa8ed8b3" (UID: "b6594e86-288e-4da2-8057-71b4fa8ed8b3"). InnerVolumeSpecName "kube-api-access-hsgwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.801227 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dfe9a33-d304-430d-90dc-86a532e34080-kube-api-access-lqgqc" (OuterVolumeSpecName: "kube-api-access-lqgqc") pod "8dfe9a33-d304-430d-90dc-86a532e34080" (UID: "8dfe9a33-d304-430d-90dc-86a532e34080"). InnerVolumeSpecName "kube-api-access-lqgqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.889121 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqgqc\" (UniqueName: \"kubernetes.io/projected/8dfe9a33-d304-430d-90dc-86a532e34080-kube-api-access-lqgqc\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:19 crc kubenswrapper[4871]: I1007 22:28:19.889172 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsgwd\" (UniqueName: \"kubernetes.io/projected/b6594e86-288e-4da2-8057-71b4fa8ed8b3-kube-api-access-hsgwd\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.124460 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0"} Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.126186 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae"} Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.126279 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c"} Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.126357 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89"} Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.126629 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f2b9-account-create-p2nlf" Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.126991 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f2b9-account-create-p2nlf" event={"ID":"8dfe9a33-d304-430d-90dc-86a532e34080","Type":"ContainerDied","Data":"ecb04e92c3c34fd869e1fc836f2f875fb05cca725a8a4b24142ca050c3f7a1ed"} Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.127074 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecb04e92c3c34fd869e1fc836f2f875fb05cca725a8a4b24142ca050c3f7a1ed" Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.128849 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6eb9-account-create-ngxgg" event={"ID":"b6594e86-288e-4da2-8057-71b4fa8ed8b3","Type":"ContainerDied","Data":"044ae24734c0f74825c0a3b867bba1e96046c1628dc5ee3d52fc40b28f9807d3"} Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.128936 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="044ae24734c0f74825c0a3b867bba1e96046c1628dc5ee3d52fc40b28f9807d3" Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.128871 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6eb9-account-create-ngxgg" Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.494416 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a79e-account-create-fwv68" Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.603903 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svm4m\" (UniqueName: \"kubernetes.io/projected/0c59e7ea-43a5-4ae5-bd0d-36bdb761a042-kube-api-access-svm4m\") pod \"0c59e7ea-43a5-4ae5-bd0d-36bdb761a042\" (UID: \"0c59e7ea-43a5-4ae5-bd0d-36bdb761a042\") " Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.622009 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c59e7ea-43a5-4ae5-bd0d-36bdb761a042-kube-api-access-svm4m" (OuterVolumeSpecName: "kube-api-access-svm4m") pod "0c59e7ea-43a5-4ae5-bd0d-36bdb761a042" (UID: "0c59e7ea-43a5-4ae5-bd0d-36bdb761a042"). InnerVolumeSpecName "kube-api-access-svm4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:20 crc kubenswrapper[4871]: I1007 22:28:20.706245 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svm4m\" (UniqueName: \"kubernetes.io/projected/0c59e7ea-43a5-4ae5-bd0d-36bdb761a042-kube-api-access-svm4m\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:21 crc kubenswrapper[4871]: I1007 22:28:21.142611 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a79e-account-create-fwv68" event={"ID":"0c59e7ea-43a5-4ae5-bd0d-36bdb761a042","Type":"ContainerDied","Data":"8759537b388dabaf81d62ee6bf88bc248bc1583ac2ae56ab6e7789ae95fbb688"} Oct 07 22:28:21 crc kubenswrapper[4871]: I1007 22:28:21.142671 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8759537b388dabaf81d62ee6bf88bc248bc1583ac2ae56ab6e7789ae95fbb688" Oct 07 22:28:21 crc kubenswrapper[4871]: I1007 22:28:21.142770 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a79e-account-create-fwv68" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.155025 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c"} Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.155433 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315"} Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.155448 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8"} Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.337476 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-tkjl5"] Oct 07 22:28:22 crc kubenswrapper[4871]: E1007 22:28:22.338233 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dfe9a33-d304-430d-90dc-86a532e34080" containerName="mariadb-account-create" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.338251 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dfe9a33-d304-430d-90dc-86a532e34080" containerName="mariadb-account-create" Oct 07 22:28:22 crc kubenswrapper[4871]: E1007 22:28:22.338282 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c59e7ea-43a5-4ae5-bd0d-36bdb761a042" containerName="mariadb-account-create" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.338290 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c59e7ea-43a5-4ae5-bd0d-36bdb761a042" containerName="mariadb-account-create" Oct 07 22:28:22 crc kubenswrapper[4871]: E1007 22:28:22.338308 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6594e86-288e-4da2-8057-71b4fa8ed8b3" containerName="mariadb-account-create" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.338315 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6594e86-288e-4da2-8057-71b4fa8ed8b3" containerName="mariadb-account-create" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.338504 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6594e86-288e-4da2-8057-71b4fa8ed8b3" containerName="mariadb-account-create" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.338531 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dfe9a33-d304-430d-90dc-86a532e34080" containerName="mariadb-account-create" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.338544 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c59e7ea-43a5-4ae5-bd0d-36bdb761a042" containerName="mariadb-account-create" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.339136 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.342218 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.343525 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8nrnc" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.348721 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tkjl5"] Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.440013 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-config-data\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.440303 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr88m\" (UniqueName: \"kubernetes.io/projected/d464df0c-28ef-4fdd-b70c-307e736def2a-kube-api-access-pr88m\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.440358 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-db-sync-config-data\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.440755 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-combined-ca-bundle\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.543387 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr88m\" (UniqueName: \"kubernetes.io/projected/d464df0c-28ef-4fdd-b70c-307e736def2a-kube-api-access-pr88m\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.543798 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-db-sync-config-data\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.544089 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-combined-ca-bundle\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.544140 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-config-data\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.550877 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-combined-ca-bundle\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.551953 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-db-sync-config-data\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.552082 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-config-data\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.564197 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr88m\" (UniqueName: \"kubernetes.io/projected/d464df0c-28ef-4fdd-b70c-307e736def2a-kube-api-access-pr88m\") pod \"glance-db-sync-tkjl5\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:22 crc kubenswrapper[4871]: I1007 22:28:22.654240 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.169177 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a"} Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.204151 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tkjl5"] Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.591432 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8vjd9" podUID="22210cb6-0be2-4f3d-9e04-d36274391d54" containerName="ovn-controller" probeResult="failure" output=< Oct 07 22:28:23 crc kubenswrapper[4871]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 07 22:28:23 crc kubenswrapper[4871]: > Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.633712 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.636124 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.882883 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8vjd9-config-dkf5c"] Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.884452 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.887273 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.903075 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8vjd9-config-dkf5c"] Oct 07 22:28:23 crc kubenswrapper[4871]: I1007 22:28:23.905254 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.077833 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run-ovn\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.077943 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-scripts\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.077970 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.078071 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q97bs\" (UniqueName: \"kubernetes.io/projected/cc3a6933-6e6c-413c-94d1-179328da1a47-kube-api-access-q97bs\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.078108 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-log-ovn\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.078190 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-additional-scripts\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.195794 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tkjl5" event={"ID":"d464df0c-28ef-4fdd-b70c-307e736def2a","Type":"ContainerStarted","Data":"c60ee4b0e6a2533df376207b7e9d032c56acdc5d4b8eb333a5a4f5d26ce72000"} Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.211788 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q97bs\" (UniqueName: \"kubernetes.io/projected/cc3a6933-6e6c-413c-94d1-179328da1a47-kube-api-access-q97bs\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.211877 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-log-ovn\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.211939 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-additional-scripts\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.212004 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run-ovn\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.212339 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-log-ovn\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.212335 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-scripts\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.212359 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run-ovn\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.212434 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.212513 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.213194 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-additional-scripts\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.214561 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-scripts\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.235136 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q97bs\" (UniqueName: \"kubernetes.io/projected/cc3a6933-6e6c-413c-94d1-179328da1a47-kube-api-access-q97bs\") pod \"ovn-controller-8vjd9-config-dkf5c\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:24 crc kubenswrapper[4871]: I1007 22:28:24.509017 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:25 crc kubenswrapper[4871]: I1007 22:28:25.044424 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8vjd9-config-dkf5c"] Oct 07 22:28:25 crc kubenswrapper[4871]: I1007 22:28:25.220073 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778"} Oct 07 22:28:25 crc kubenswrapper[4871]: I1007 22:28:25.220502 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a"} Oct 07 22:28:25 crc kubenswrapper[4871]: I1007 22:28:25.220517 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c"} Oct 07 22:28:25 crc kubenswrapper[4871]: I1007 22:28:25.220529 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057"} Oct 07 22:28:25 crc kubenswrapper[4871]: I1007 22:28:25.221682 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8vjd9-config-dkf5c" event={"ID":"cc3a6933-6e6c-413c-94d1-179328da1a47","Type":"ContainerStarted","Data":"4a18a12d011423d1dc00614b9f9a8ee6c14ac23c4851ef65b0056b49d3cd0cbe"} Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.245243 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad"} Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.245293 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9"} Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.245303 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerStarted","Data":"5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d"} Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.251635 4871 generic.go:334] "Generic (PLEG): container finished" podID="cc3a6933-6e6c-413c-94d1-179328da1a47" containerID="52ca931b6e665fac4e12d71a04df651d6da1645c3ce0598832d89f347ee1e924" exitCode=0 Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.251677 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8vjd9-config-dkf5c" event={"ID":"cc3a6933-6e6c-413c-94d1-179328da1a47","Type":"ContainerDied","Data":"52ca931b6e665fac4e12d71a04df651d6da1645c3ce0598832d89f347ee1e924"} Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.289667 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.760485096 podStartE2EDuration="28.289636505s" podCreationTimestamp="2025-10-07 22:27:58 +0000 UTC" firstStartedPulling="2025-10-07 22:28:16.76537579 +0000 UTC m=+1170.568073883" lastFinishedPulling="2025-10-07 22:28:24.294527219 +0000 UTC m=+1178.097225292" observedRunningTime="2025-10-07 22:28:26.284649783 +0000 UTC m=+1180.087347886" watchObservedRunningTime="2025-10-07 22:28:26.289636505 +0000 UTC m=+1180.092334588" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.578393 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-f26k9"] Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.580157 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.584437 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.594881 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-f26k9"] Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.759759 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-config\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.759831 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-sb\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.759912 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-nb\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.759936 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-swift-storage-0\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.759960 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-svc\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.759997 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m452\" (UniqueName: \"kubernetes.io/projected/63e41063-fa11-4824-87df-49c82c6d9921-kube-api-access-7m452\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.861037 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-config\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.861086 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-sb\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.861169 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-nb\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.861192 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-swift-storage-0\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.861217 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-svc\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.861256 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m452\" (UniqueName: \"kubernetes.io/projected/63e41063-fa11-4824-87df-49c82c6d9921-kube-api-access-7m452\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.862137 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-sb\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.862354 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-swift-storage-0\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.862618 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-nb\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.862635 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-svc\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.863016 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-config\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.887992 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m452\" (UniqueName: \"kubernetes.io/projected/63e41063-fa11-4824-87df-49c82c6d9921-kube-api-access-7m452\") pod \"dnsmasq-dns-59f45f6cf7-f26k9\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:26 crc kubenswrapper[4871]: I1007 22:28:26.906442 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.374906 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-f26k9"] Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.589538 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.778618 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run\") pod \"cc3a6933-6e6c-413c-94d1-179328da1a47\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.778696 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-additional-scripts\") pod \"cc3a6933-6e6c-413c-94d1-179328da1a47\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.778722 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-scripts\") pod \"cc3a6933-6e6c-413c-94d1-179328da1a47\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.778764 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run" (OuterVolumeSpecName: "var-run") pod "cc3a6933-6e6c-413c-94d1-179328da1a47" (UID: "cc3a6933-6e6c-413c-94d1-179328da1a47"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.778785 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run-ovn\") pod \"cc3a6933-6e6c-413c-94d1-179328da1a47\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.778818 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-log-ovn\") pod \"cc3a6933-6e6c-413c-94d1-179328da1a47\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.778975 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q97bs\" (UniqueName: \"kubernetes.io/projected/cc3a6933-6e6c-413c-94d1-179328da1a47-kube-api-access-q97bs\") pod \"cc3a6933-6e6c-413c-94d1-179328da1a47\" (UID: \"cc3a6933-6e6c-413c-94d1-179328da1a47\") " Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.779113 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "cc3a6933-6e6c-413c-94d1-179328da1a47" (UID: "cc3a6933-6e6c-413c-94d1-179328da1a47"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.779164 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "cc3a6933-6e6c-413c-94d1-179328da1a47" (UID: "cc3a6933-6e6c-413c-94d1-179328da1a47"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.779357 4871 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.779376 4871 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.779386 4871 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cc3a6933-6e6c-413c-94d1-179328da1a47-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.779826 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "cc3a6933-6e6c-413c-94d1-179328da1a47" (UID: "cc3a6933-6e6c-413c-94d1-179328da1a47"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.780074 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-scripts" (OuterVolumeSpecName: "scripts") pod "cc3a6933-6e6c-413c-94d1-179328da1a47" (UID: "cc3a6933-6e6c-413c-94d1-179328da1a47"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.783851 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc3a6933-6e6c-413c-94d1-179328da1a47-kube-api-access-q97bs" (OuterVolumeSpecName: "kube-api-access-q97bs") pod "cc3a6933-6e6c-413c-94d1-179328da1a47" (UID: "cc3a6933-6e6c-413c-94d1-179328da1a47"). InnerVolumeSpecName "kube-api-access-q97bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.880966 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q97bs\" (UniqueName: \"kubernetes.io/projected/cc3a6933-6e6c-413c-94d1-179328da1a47-kube-api-access-q97bs\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.880997 4871 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:27 crc kubenswrapper[4871]: I1007 22:28:27.881008 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc3a6933-6e6c-413c-94d1-179328da1a47-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.277019 4871 generic.go:334] "Generic (PLEG): container finished" podID="63e41063-fa11-4824-87df-49c82c6d9921" containerID="4ef6f7552387c8d751270f4aee8227aba0784289050826a18691cd405931e855" exitCode=0 Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.277208 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" event={"ID":"63e41063-fa11-4824-87df-49c82c6d9921","Type":"ContainerDied","Data":"4ef6f7552387c8d751270f4aee8227aba0784289050826a18691cd405931e855"} Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.277378 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" event={"ID":"63e41063-fa11-4824-87df-49c82c6d9921","Type":"ContainerStarted","Data":"cc76a46d042345982a49ccdcac3ea839c0b32425bddd8f66ae983ca2aecfecee"} Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.285991 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8vjd9-config-dkf5c" event={"ID":"cc3a6933-6e6c-413c-94d1-179328da1a47","Type":"ContainerDied","Data":"4a18a12d011423d1dc00614b9f9a8ee6c14ac23c4851ef65b0056b49d3cd0cbe"} Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.286042 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a18a12d011423d1dc00614b9f9a8ee6c14ac23c4851ef65b0056b49d3cd0cbe" Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.286097 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8vjd9-config-dkf5c" Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.590661 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-8vjd9" Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.701262 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8vjd9-config-dkf5c"] Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.708395 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8vjd9-config-dkf5c"] Oct 07 22:28:28 crc kubenswrapper[4871]: I1007 22:28:28.992569 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc3a6933-6e6c-413c-94d1-179328da1a47" path="/var/lib/kubelet/pods/cc3a6933-6e6c-413c-94d1-179328da1a47/volumes" Oct 07 22:28:29 crc kubenswrapper[4871]: I1007 22:28:29.299310 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" event={"ID":"63e41063-fa11-4824-87df-49c82c6d9921","Type":"ContainerStarted","Data":"02fb7a6e347fb3a558e6cd6a6dbef20648452461f42bf80aff0be2b8bfc262be"} Oct 07 22:28:29 crc kubenswrapper[4871]: I1007 22:28:29.299514 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:29 crc kubenswrapper[4871]: I1007 22:28:29.340198 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" podStartSLOduration=3.340165483 podStartE2EDuration="3.340165483s" podCreationTimestamp="2025-10-07 22:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:29.32192132 +0000 UTC m=+1183.124619403" watchObservedRunningTime="2025-10-07 22:28:29.340165483 +0000 UTC m=+1183.142863556" Oct 07 22:28:33 crc kubenswrapper[4871]: I1007 22:28:33.909051 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.222143 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.262702 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-stzkd"] Oct 07 22:28:34 crc kubenswrapper[4871]: E1007 22:28:34.263070 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3a6933-6e6c-413c-94d1-179328da1a47" containerName="ovn-config" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.263088 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3a6933-6e6c-413c-94d1-179328da1a47" containerName="ovn-config" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.263276 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3a6933-6e6c-413c-94d1-179328da1a47" containerName="ovn-config" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.264782 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-stzkd" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.300117 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-stzkd"] Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.322140 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkc5d\" (UniqueName: \"kubernetes.io/projected/ce3ca1ea-9db5-4974-b974-ba4d25bf8e09-kube-api-access-kkc5d\") pod \"barbican-db-create-stzkd\" (UID: \"ce3ca1ea-9db5-4974-b974-ba4d25bf8e09\") " pod="openstack/barbican-db-create-stzkd" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.423600 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkc5d\" (UniqueName: \"kubernetes.io/projected/ce3ca1ea-9db5-4974-b974-ba4d25bf8e09-kube-api-access-kkc5d\") pod \"barbican-db-create-stzkd\" (UID: \"ce3ca1ea-9db5-4974-b974-ba4d25bf8e09\") " pod="openstack/barbican-db-create-stzkd" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.460094 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkc5d\" (UniqueName: \"kubernetes.io/projected/ce3ca1ea-9db5-4974-b974-ba4d25bf8e09-kube-api-access-kkc5d\") pod \"barbican-db-create-stzkd\" (UID: \"ce3ca1ea-9db5-4974-b974-ba4d25bf8e09\") " pod="openstack/barbican-db-create-stzkd" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.501116 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-b2s89"] Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.502311 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b2s89" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.528090 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b2s89"] Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.587996 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-stzkd" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.621548 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-s2nds"] Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.624402 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s2nds" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.626609 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k77br\" (UniqueName: \"kubernetes.io/projected/b4f0df77-5bfb-4658-9349-b5089da279d8-kube-api-access-k77br\") pod \"cinder-db-create-b2s89\" (UID: \"b4f0df77-5bfb-4658-9349-b5089da279d8\") " pod="openstack/cinder-db-create-b2s89" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.641979 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s2nds"] Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.697460 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xt474"] Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.699330 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.705403 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.705599 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.706174 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4kmth" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.706321 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.717121 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xt474"] Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.729303 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlkq2\" (UniqueName: \"kubernetes.io/projected/b78d2643-0d33-49d1-9230-64c980990c7b-kube-api-access-nlkq2\") pod \"neutron-db-create-s2nds\" (UID: \"b78d2643-0d33-49d1-9230-64c980990c7b\") " pod="openstack/neutron-db-create-s2nds" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.729790 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k77br\" (UniqueName: \"kubernetes.io/projected/b4f0df77-5bfb-4658-9349-b5089da279d8-kube-api-access-k77br\") pod \"cinder-db-create-b2s89\" (UID: \"b4f0df77-5bfb-4658-9349-b5089da279d8\") " pod="openstack/cinder-db-create-b2s89" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.752726 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k77br\" (UniqueName: \"kubernetes.io/projected/b4f0df77-5bfb-4658-9349-b5089da279d8-kube-api-access-k77br\") pod \"cinder-db-create-b2s89\" (UID: \"b4f0df77-5bfb-4658-9349-b5089da279d8\") " pod="openstack/cinder-db-create-b2s89" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.832303 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlkq2\" (UniqueName: \"kubernetes.io/projected/b78d2643-0d33-49d1-9230-64c980990c7b-kube-api-access-nlkq2\") pod \"neutron-db-create-s2nds\" (UID: \"b78d2643-0d33-49d1-9230-64c980990c7b\") " pod="openstack/neutron-db-create-s2nds" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.832390 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-config-data\") pod \"keystone-db-sync-xt474\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.832432 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p9f2\" (UniqueName: \"kubernetes.io/projected/f23add1a-b7b7-4670-b510-985a7c4c221d-kube-api-access-6p9f2\") pod \"keystone-db-sync-xt474\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.832490 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-combined-ca-bundle\") pod \"keystone-db-sync-xt474\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.853681 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlkq2\" (UniqueName: \"kubernetes.io/projected/b78d2643-0d33-49d1-9230-64c980990c7b-kube-api-access-nlkq2\") pod \"neutron-db-create-s2nds\" (UID: \"b78d2643-0d33-49d1-9230-64c980990c7b\") " pod="openstack/neutron-db-create-s2nds" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.874195 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b2s89" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.934025 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-config-data\") pod \"keystone-db-sync-xt474\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.934079 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p9f2\" (UniqueName: \"kubernetes.io/projected/f23add1a-b7b7-4670-b510-985a7c4c221d-kube-api-access-6p9f2\") pod \"keystone-db-sync-xt474\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.934130 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-combined-ca-bundle\") pod \"keystone-db-sync-xt474\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.938472 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-combined-ca-bundle\") pod \"keystone-db-sync-xt474\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.938893 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-config-data\") pod \"keystone-db-sync-xt474\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.954344 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p9f2\" (UniqueName: \"kubernetes.io/projected/f23add1a-b7b7-4670-b510-985a7c4c221d-kube-api-access-6p9f2\") pod \"keystone-db-sync-xt474\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:34 crc kubenswrapper[4871]: I1007 22:28:34.958346 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s2nds" Oct 07 22:28:35 crc kubenswrapper[4871]: I1007 22:28:35.029132 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:36 crc kubenswrapper[4871]: I1007 22:28:36.908035 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:28:37 crc kubenswrapper[4871]: I1007 22:28:37.022045 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-qg2kn"] Oct 07 22:28:37 crc kubenswrapper[4871]: I1007 22:28:37.022382 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" podUID="82f27d48-b207-4c5d-a981-102a64909f5e" containerName="dnsmasq-dns" containerID="cri-o://8b399b7d37e73b596db12fe18565964c2f91556cbf97326d9e702af147df24d6" gracePeriod=10 Oct 07 22:28:37 crc kubenswrapper[4871]: I1007 22:28:37.394096 4871 generic.go:334] "Generic (PLEG): container finished" podID="82f27d48-b207-4c5d-a981-102a64909f5e" containerID="8b399b7d37e73b596db12fe18565964c2f91556cbf97326d9e702af147df24d6" exitCode=0 Oct 07 22:28:37 crc kubenswrapper[4871]: I1007 22:28:37.394141 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" event={"ID":"82f27d48-b207-4c5d-a981-102a64909f5e","Type":"ContainerDied","Data":"8b399b7d37e73b596db12fe18565964c2f91556cbf97326d9e702af147df24d6"} Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.373864 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.433635 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.433986 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" event={"ID":"82f27d48-b207-4c5d-a981-102a64909f5e","Type":"ContainerDied","Data":"697158c8bd7356ef549520d0934aaddc6ee7119f610b5720f943fc169ffc17d9"} Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.434060 4871 scope.go:117] "RemoveContainer" containerID="8b399b7d37e73b596db12fe18565964c2f91556cbf97326d9e702af147df24d6" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.438646 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-nb\") pod \"82f27d48-b207-4c5d-a981-102a64909f5e\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.438692 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-sb\") pod \"82f27d48-b207-4c5d-a981-102a64909f5e\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.438718 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-config\") pod \"82f27d48-b207-4c5d-a981-102a64909f5e\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.438762 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-dns-svc\") pod \"82f27d48-b207-4c5d-a981-102a64909f5e\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.438818 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc7np\" (UniqueName: \"kubernetes.io/projected/82f27d48-b207-4c5d-a981-102a64909f5e-kube-api-access-zc7np\") pod \"82f27d48-b207-4c5d-a981-102a64909f5e\" (UID: \"82f27d48-b207-4c5d-a981-102a64909f5e\") " Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.460728 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82f27d48-b207-4c5d-a981-102a64909f5e-kube-api-access-zc7np" (OuterVolumeSpecName: "kube-api-access-zc7np") pod "82f27d48-b207-4c5d-a981-102a64909f5e" (UID: "82f27d48-b207-4c5d-a981-102a64909f5e"). InnerVolumeSpecName "kube-api-access-zc7np". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.481285 4871 scope.go:117] "RemoveContainer" containerID="3490a71fe3931fb507ef1f0808f6041cb516b64c97f3f38d2cfa61e0a7b4398f" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.499776 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "82f27d48-b207-4c5d-a981-102a64909f5e" (UID: "82f27d48-b207-4c5d-a981-102a64909f5e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.530894 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-config" (OuterVolumeSpecName: "config") pod "82f27d48-b207-4c5d-a981-102a64909f5e" (UID: "82f27d48-b207-4c5d-a981-102a64909f5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.541305 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.541337 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.541347 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc7np\" (UniqueName: \"kubernetes.io/projected/82f27d48-b207-4c5d-a981-102a64909f5e-kube-api-access-zc7np\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.543717 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "82f27d48-b207-4c5d-a981-102a64909f5e" (UID: "82f27d48-b207-4c5d-a981-102a64909f5e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.556271 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "82f27d48-b207-4c5d-a981-102a64909f5e" (UID: "82f27d48-b207-4c5d-a981-102a64909f5e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.643502 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.643534 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82f27d48-b207-4c5d-a981-102a64909f5e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.706485 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b2s89"] Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.721128 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xt474"] Oct 07 22:28:39 crc kubenswrapper[4871]: W1007 22:28:39.722271 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4f0df77_5bfb_4658_9349_b5089da279d8.slice/crio-25ea1b6263506fb67bcc6e5502fade88a14909f8e397686003b45a8944a4a668 WatchSource:0}: Error finding container 25ea1b6263506fb67bcc6e5502fade88a14909f8e397686003b45a8944a4a668: Status 404 returned error can't find the container with id 25ea1b6263506fb67bcc6e5502fade88a14909f8e397686003b45a8944a4a668 Oct 07 22:28:39 crc kubenswrapper[4871]: W1007 22:28:39.727235 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf23add1a_b7b7_4670_b510_985a7c4c221d.slice/crio-5070f880cc4682041223445a5350f2b182030d511e17989be4de4290a91af254 WatchSource:0}: Error finding container 5070f880cc4682041223445a5350f2b182030d511e17989be4de4290a91af254: Status 404 returned error can't find the container with id 5070f880cc4682041223445a5350f2b182030d511e17989be4de4290a91af254 Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.777689 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-qg2kn"] Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.791157 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-qg2kn"] Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.850920 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s2nds"] Oct 07 22:28:39 crc kubenswrapper[4871]: I1007 22:28:39.889812 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-stzkd"] Oct 07 22:28:39 crc kubenswrapper[4871]: W1007 22:28:39.892522 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce3ca1ea_9db5_4974_b974_ba4d25bf8e09.slice/crio-fd26d350bf87189696610283467572e3150aa55e4746f39dc02033abcb5d9234 WatchSource:0}: Error finding container fd26d350bf87189696610283467572e3150aa55e4746f39dc02033abcb5d9234: Status 404 returned error can't find the container with id fd26d350bf87189696610283467572e3150aa55e4746f39dc02033abcb5d9234 Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.441198 4871 generic.go:334] "Generic (PLEG): container finished" podID="b78d2643-0d33-49d1-9230-64c980990c7b" containerID="7da0fc02050c5905620a8024d8e1e85e13be4ceca818e33fb1a1a757f43a3f91" exitCode=0 Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.441422 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s2nds" event={"ID":"b78d2643-0d33-49d1-9230-64c980990c7b","Type":"ContainerDied","Data":"7da0fc02050c5905620a8024d8e1e85e13be4ceca818e33fb1a1a757f43a3f91"} Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.441649 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s2nds" event={"ID":"b78d2643-0d33-49d1-9230-64c980990c7b","Type":"ContainerStarted","Data":"a0b15662b3a7efc6f6f6bac9f5256f8e1e3010b45881bd9ac340be511b997d3b"} Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.452856 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tkjl5" event={"ID":"d464df0c-28ef-4fdd-b70c-307e736def2a","Type":"ContainerStarted","Data":"4a7a0515437b353511dce4852c471123dc48a9388970a160c88a5a0ae9a250d1"} Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.461178 4871 generic.go:334] "Generic (PLEG): container finished" podID="ce3ca1ea-9db5-4974-b974-ba4d25bf8e09" containerID="4a3cd78bb6c3e29ad022309e0781d2d390b88fc52213fdb0aae6d15cca708568" exitCode=0 Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.461279 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-stzkd" event={"ID":"ce3ca1ea-9db5-4974-b974-ba4d25bf8e09","Type":"ContainerDied","Data":"4a3cd78bb6c3e29ad022309e0781d2d390b88fc52213fdb0aae6d15cca708568"} Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.461327 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-stzkd" event={"ID":"ce3ca1ea-9db5-4974-b974-ba4d25bf8e09","Type":"ContainerStarted","Data":"fd26d350bf87189696610283467572e3150aa55e4746f39dc02033abcb5d9234"} Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.473454 4871 generic.go:334] "Generic (PLEG): container finished" podID="b4f0df77-5bfb-4658-9349-b5089da279d8" containerID="d0224dd10a0847f3e3b8e29d449dcabf476a8b5f0e8ee506885df65b41d807f8" exitCode=0 Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.473533 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b2s89" event={"ID":"b4f0df77-5bfb-4658-9349-b5089da279d8","Type":"ContainerDied","Data":"d0224dd10a0847f3e3b8e29d449dcabf476a8b5f0e8ee506885df65b41d807f8"} Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.473570 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b2s89" event={"ID":"b4f0df77-5bfb-4658-9349-b5089da279d8","Type":"ContainerStarted","Data":"25ea1b6263506fb67bcc6e5502fade88a14909f8e397686003b45a8944a4a668"} Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.476589 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xt474" event={"ID":"f23add1a-b7b7-4670-b510-985a7c4c221d","Type":"ContainerStarted","Data":"5070f880cc4682041223445a5350f2b182030d511e17989be4de4290a91af254"} Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.506622 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-tkjl5" podStartSLOduration=2.534656545 podStartE2EDuration="18.506594678s" podCreationTimestamp="2025-10-07 22:28:22 +0000 UTC" firstStartedPulling="2025-10-07 22:28:23.217326509 +0000 UTC m=+1177.020024592" lastFinishedPulling="2025-10-07 22:28:39.189264662 +0000 UTC m=+1192.991962725" observedRunningTime="2025-10-07 22:28:40.505632353 +0000 UTC m=+1194.308330626" watchObservedRunningTime="2025-10-07 22:28:40.506594678 +0000 UTC m=+1194.309292761" Oct 07 22:28:40 crc kubenswrapper[4871]: I1007 22:28:40.997073 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82f27d48-b207-4c5d-a981-102a64909f5e" path="/var/lib/kubelet/pods/82f27d48-b207-4c5d-a981-102a64909f5e/volumes" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.006202 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b2s89" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.013841 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s2nds" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.025482 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-stzkd" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.100344 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkc5d\" (UniqueName: \"kubernetes.io/projected/ce3ca1ea-9db5-4974-b974-ba4d25bf8e09-kube-api-access-kkc5d\") pod \"ce3ca1ea-9db5-4974-b974-ba4d25bf8e09\" (UID: \"ce3ca1ea-9db5-4974-b974-ba4d25bf8e09\") " Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.100441 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k77br\" (UniqueName: \"kubernetes.io/projected/b4f0df77-5bfb-4658-9349-b5089da279d8-kube-api-access-k77br\") pod \"b4f0df77-5bfb-4658-9349-b5089da279d8\" (UID: \"b4f0df77-5bfb-4658-9349-b5089da279d8\") " Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.100520 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlkq2\" (UniqueName: \"kubernetes.io/projected/b78d2643-0d33-49d1-9230-64c980990c7b-kube-api-access-nlkq2\") pod \"b78d2643-0d33-49d1-9230-64c980990c7b\" (UID: \"b78d2643-0d33-49d1-9230-64c980990c7b\") " Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.110437 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce3ca1ea-9db5-4974-b974-ba4d25bf8e09-kube-api-access-kkc5d" (OuterVolumeSpecName: "kube-api-access-kkc5d") pod "ce3ca1ea-9db5-4974-b974-ba4d25bf8e09" (UID: "ce3ca1ea-9db5-4974-b974-ba4d25bf8e09"). InnerVolumeSpecName "kube-api-access-kkc5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.110492 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b78d2643-0d33-49d1-9230-64c980990c7b-kube-api-access-nlkq2" (OuterVolumeSpecName: "kube-api-access-nlkq2") pod "b78d2643-0d33-49d1-9230-64c980990c7b" (UID: "b78d2643-0d33-49d1-9230-64c980990c7b"). InnerVolumeSpecName "kube-api-access-nlkq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.112224 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4f0df77-5bfb-4658-9349-b5089da279d8-kube-api-access-k77br" (OuterVolumeSpecName: "kube-api-access-k77br") pod "b4f0df77-5bfb-4658-9349-b5089da279d8" (UID: "b4f0df77-5bfb-4658-9349-b5089da279d8"). InnerVolumeSpecName "kube-api-access-k77br". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.203126 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k77br\" (UniqueName: \"kubernetes.io/projected/b4f0df77-5bfb-4658-9349-b5089da279d8-kube-api-access-k77br\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.203180 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlkq2\" (UniqueName: \"kubernetes.io/projected/b78d2643-0d33-49d1-9230-64c980990c7b-kube-api-access-nlkq2\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.203193 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkc5d\" (UniqueName: \"kubernetes.io/projected/ce3ca1ea-9db5-4974-b974-ba4d25bf8e09-kube-api-access-kkc5d\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.508535 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b2s89" event={"ID":"b4f0df77-5bfb-4658-9349-b5089da279d8","Type":"ContainerDied","Data":"25ea1b6263506fb67bcc6e5502fade88a14909f8e397686003b45a8944a4a668"} Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.508598 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25ea1b6263506fb67bcc6e5502fade88a14909f8e397686003b45a8944a4a668" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.508678 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b2s89" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.510570 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s2nds" event={"ID":"b78d2643-0d33-49d1-9230-64c980990c7b","Type":"ContainerDied","Data":"a0b15662b3a7efc6f6f6bac9f5256f8e1e3010b45881bd9ac340be511b997d3b"} Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.510605 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0b15662b3a7efc6f6f6bac9f5256f8e1e3010b45881bd9ac340be511b997d3b" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.510654 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s2nds" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.519387 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-stzkd" event={"ID":"ce3ca1ea-9db5-4974-b974-ba4d25bf8e09","Type":"ContainerDied","Data":"fd26d350bf87189696610283467572e3150aa55e4746f39dc02033abcb5d9234"} Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.519452 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd26d350bf87189696610283467572e3150aa55e4746f39dc02033abcb5d9234" Oct 07 22:28:42 crc kubenswrapper[4871]: I1007 22:28:42.519567 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-stzkd" Oct 07 22:28:44 crc kubenswrapper[4871]: I1007 22:28:44.245664 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57f58c7cff-qg2kn" podUID="82f27d48-b207-4c5d-a981-102a64909f5e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.118:5353: i/o timeout" Oct 07 22:28:47 crc kubenswrapper[4871]: I1007 22:28:47.583543 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xt474" event={"ID":"f23add1a-b7b7-4670-b510-985a7c4c221d","Type":"ContainerStarted","Data":"c977ba81193a781589b8de9336bdffb902f2bc69bde7a885590c026fb068065b"} Oct 07 22:28:47 crc kubenswrapper[4871]: I1007 22:28:47.637201 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xt474" podStartSLOduration=6.761103845 podStartE2EDuration="13.63717795s" podCreationTimestamp="2025-10-07 22:28:34 +0000 UTC" firstStartedPulling="2025-10-07 22:28:39.733508811 +0000 UTC m=+1193.536206884" lastFinishedPulling="2025-10-07 22:28:46.609582916 +0000 UTC m=+1200.412280989" observedRunningTime="2025-10-07 22:28:47.620470178 +0000 UTC m=+1201.423168251" watchObservedRunningTime="2025-10-07 22:28:47.63717795 +0000 UTC m=+1201.439876024" Oct 07 22:28:48 crc kubenswrapper[4871]: I1007 22:28:48.604612 4871 generic.go:334] "Generic (PLEG): container finished" podID="d464df0c-28ef-4fdd-b70c-307e736def2a" containerID="4a7a0515437b353511dce4852c471123dc48a9388970a160c88a5a0ae9a250d1" exitCode=0 Oct 07 22:28:48 crc kubenswrapper[4871]: I1007 22:28:48.604688 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tkjl5" event={"ID":"d464df0c-28ef-4fdd-b70c-307e736def2a","Type":"ContainerDied","Data":"4a7a0515437b353511dce4852c471123dc48a9388970a160c88a5a0ae9a250d1"} Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.084222 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.178057 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-db-sync-config-data\") pod \"d464df0c-28ef-4fdd-b70c-307e736def2a\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.178736 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-combined-ca-bundle\") pod \"d464df0c-28ef-4fdd-b70c-307e736def2a\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.179030 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-config-data\") pod \"d464df0c-28ef-4fdd-b70c-307e736def2a\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.179198 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr88m\" (UniqueName: \"kubernetes.io/projected/d464df0c-28ef-4fdd-b70c-307e736def2a-kube-api-access-pr88m\") pod \"d464df0c-28ef-4fdd-b70c-307e736def2a\" (UID: \"d464df0c-28ef-4fdd-b70c-307e736def2a\") " Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.186773 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d464df0c-28ef-4fdd-b70c-307e736def2a" (UID: "d464df0c-28ef-4fdd-b70c-307e736def2a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.190091 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d464df0c-28ef-4fdd-b70c-307e736def2a-kube-api-access-pr88m" (OuterVolumeSpecName: "kube-api-access-pr88m") pod "d464df0c-28ef-4fdd-b70c-307e736def2a" (UID: "d464df0c-28ef-4fdd-b70c-307e736def2a"). InnerVolumeSpecName "kube-api-access-pr88m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.245682 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d464df0c-28ef-4fdd-b70c-307e736def2a" (UID: "d464df0c-28ef-4fdd-b70c-307e736def2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.248669 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-config-data" (OuterVolumeSpecName: "config-data") pod "d464df0c-28ef-4fdd-b70c-307e736def2a" (UID: "d464df0c-28ef-4fdd-b70c-307e736def2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.282327 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.282586 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr88m\" (UniqueName: \"kubernetes.io/projected/d464df0c-28ef-4fdd-b70c-307e736def2a-kube-api-access-pr88m\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.282719 4871 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.282825 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d464df0c-28ef-4fdd-b70c-307e736def2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.625602 4871 generic.go:334] "Generic (PLEG): container finished" podID="f23add1a-b7b7-4670-b510-985a7c4c221d" containerID="c977ba81193a781589b8de9336bdffb902f2bc69bde7a885590c026fb068065b" exitCode=0 Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.625722 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xt474" event={"ID":"f23add1a-b7b7-4670-b510-985a7c4c221d","Type":"ContainerDied","Data":"c977ba81193a781589b8de9336bdffb902f2bc69bde7a885590c026fb068065b"} Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.629980 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tkjl5" event={"ID":"d464df0c-28ef-4fdd-b70c-307e736def2a","Type":"ContainerDied","Data":"c60ee4b0e6a2533df376207b7e9d032c56acdc5d4b8eb333a5a4f5d26ce72000"} Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.630025 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c60ee4b0e6a2533df376207b7e9d032c56acdc5d4b8eb333a5a4f5d26ce72000" Oct 07 22:28:50 crc kubenswrapper[4871]: I1007 22:28:50.630124 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tkjl5" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.072551 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf6f78f57-x7clx"] Oct 07 22:28:51 crc kubenswrapper[4871]: E1007 22:28:51.072913 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f27d48-b207-4c5d-a981-102a64909f5e" containerName="dnsmasq-dns" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.072932 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f27d48-b207-4c5d-a981-102a64909f5e" containerName="dnsmasq-dns" Oct 07 22:28:51 crc kubenswrapper[4871]: E1007 22:28:51.072952 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f0df77-5bfb-4658-9349-b5089da279d8" containerName="mariadb-database-create" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.072958 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f0df77-5bfb-4658-9349-b5089da279d8" containerName="mariadb-database-create" Oct 07 22:28:51 crc kubenswrapper[4871]: E1007 22:28:51.072968 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d464df0c-28ef-4fdd-b70c-307e736def2a" containerName="glance-db-sync" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.072976 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d464df0c-28ef-4fdd-b70c-307e736def2a" containerName="glance-db-sync" Oct 07 22:28:51 crc kubenswrapper[4871]: E1007 22:28:51.072991 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3ca1ea-9db5-4974-b974-ba4d25bf8e09" containerName="mariadb-database-create" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.072997 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3ca1ea-9db5-4974-b974-ba4d25bf8e09" containerName="mariadb-database-create" Oct 07 22:28:51 crc kubenswrapper[4871]: E1007 22:28:51.073012 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f27d48-b207-4c5d-a981-102a64909f5e" containerName="init" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.073017 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f27d48-b207-4c5d-a981-102a64909f5e" containerName="init" Oct 07 22:28:51 crc kubenswrapper[4871]: E1007 22:28:51.073038 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78d2643-0d33-49d1-9230-64c980990c7b" containerName="mariadb-database-create" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.073044 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78d2643-0d33-49d1-9230-64c980990c7b" containerName="mariadb-database-create" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.073191 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="82f27d48-b207-4c5d-a981-102a64909f5e" containerName="dnsmasq-dns" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.073203 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3ca1ea-9db5-4974-b974-ba4d25bf8e09" containerName="mariadb-database-create" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.073224 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4f0df77-5bfb-4658-9349-b5089da279d8" containerName="mariadb-database-create" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.073236 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d464df0c-28ef-4fdd-b70c-307e736def2a" containerName="glance-db-sync" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.073245 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78d2643-0d33-49d1-9230-64c980990c7b" containerName="mariadb-database-create" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.074116 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.098755 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf6f78f57-x7clx"] Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.201681 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-svc\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.201760 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-config\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.201953 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpfp7\" (UniqueName: \"kubernetes.io/projected/eaba3b8a-2131-463a-9dae-1454979ced8b-kube-api-access-jpfp7\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.202020 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-sb\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.202043 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-swift-storage-0\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.202077 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-nb\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.303274 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpfp7\" (UniqueName: \"kubernetes.io/projected/eaba3b8a-2131-463a-9dae-1454979ced8b-kube-api-access-jpfp7\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.303365 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-sb\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.303395 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-swift-storage-0\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.303428 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-nb\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.303468 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-config\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.303490 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-svc\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.304645 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-svc\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.304638 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-sb\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.304645 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-config\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.304917 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-nb\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.305085 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-swift-storage-0\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.325199 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpfp7\" (UniqueName: \"kubernetes.io/projected/eaba3b8a-2131-463a-9dae-1454979ced8b-kube-api-access-jpfp7\") pod \"dnsmasq-dns-cf6f78f57-x7clx\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.397039 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:51 crc kubenswrapper[4871]: I1007 22:28:51.733681 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf6f78f57-x7clx"] Oct 07 22:28:51 crc kubenswrapper[4871]: W1007 22:28:51.753011 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeaba3b8a_2131_463a_9dae_1454979ced8b.slice/crio-a0bc0a1485e7fea4de23737f1d61e6300a251e146592e2b44ac64f778bb0be0c WatchSource:0}: Error finding container a0bc0a1485e7fea4de23737f1d61e6300a251e146592e2b44ac64f778bb0be0c: Status 404 returned error can't find the container with id a0bc0a1485e7fea4de23737f1d61e6300a251e146592e2b44ac64f778bb0be0c Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:51.948060 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.138517 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-combined-ca-bundle\") pod \"f23add1a-b7b7-4670-b510-985a7c4c221d\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.138598 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p9f2\" (UniqueName: \"kubernetes.io/projected/f23add1a-b7b7-4670-b510-985a7c4c221d-kube-api-access-6p9f2\") pod \"f23add1a-b7b7-4670-b510-985a7c4c221d\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.138836 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-config-data\") pod \"f23add1a-b7b7-4670-b510-985a7c4c221d\" (UID: \"f23add1a-b7b7-4670-b510-985a7c4c221d\") " Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.146933 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f23add1a-b7b7-4670-b510-985a7c4c221d-kube-api-access-6p9f2" (OuterVolumeSpecName: "kube-api-access-6p9f2") pod "f23add1a-b7b7-4670-b510-985a7c4c221d" (UID: "f23add1a-b7b7-4670-b510-985a7c4c221d"). InnerVolumeSpecName "kube-api-access-6p9f2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.176937 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f23add1a-b7b7-4670-b510-985a7c4c221d" (UID: "f23add1a-b7b7-4670-b510-985a7c4c221d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.191386 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-config-data" (OuterVolumeSpecName: "config-data") pod "f23add1a-b7b7-4670-b510-985a7c4c221d" (UID: "f23add1a-b7b7-4670-b510-985a7c4c221d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.240724 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p9f2\" (UniqueName: \"kubernetes.io/projected/f23add1a-b7b7-4670-b510-985a7c4c221d-kube-api-access-6p9f2\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.240759 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.240773 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23add1a-b7b7-4670-b510-985a7c4c221d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.653324 4871 generic.go:334] "Generic (PLEG): container finished" podID="eaba3b8a-2131-463a-9dae-1454979ced8b" containerID="115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901" exitCode=0 Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.653498 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" event={"ID":"eaba3b8a-2131-463a-9dae-1454979ced8b","Type":"ContainerDied","Data":"115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901"} Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.653856 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" event={"ID":"eaba3b8a-2131-463a-9dae-1454979ced8b","Type":"ContainerStarted","Data":"a0bc0a1485e7fea4de23737f1d61e6300a251e146592e2b44ac64f778bb0be0c"} Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.660399 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xt474" event={"ID":"f23add1a-b7b7-4670-b510-985a7c4c221d","Type":"ContainerDied","Data":"5070f880cc4682041223445a5350f2b182030d511e17989be4de4290a91af254"} Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.660458 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5070f880cc4682041223445a5350f2b182030d511e17989be4de4290a91af254" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.660504 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xt474" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.878309 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-b47pf"] Oct 07 22:28:52 crc kubenswrapper[4871]: E1007 22:28:52.878681 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f23add1a-b7b7-4670-b510-985a7c4c221d" containerName="keystone-db-sync" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.878697 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f23add1a-b7b7-4670-b510-985a7c4c221d" containerName="keystone-db-sync" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.878886 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f23add1a-b7b7-4670-b510-985a7c4c221d" containerName="keystone-db-sync" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.879471 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.881575 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.881743 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.881745 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4kmth" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.881914 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.894673 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b47pf"] Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.915861 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf6f78f57-x7clx"] Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.948273 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8d44b7457-sgd4w"] Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.949640 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:52 crc kubenswrapper[4871]: I1007 22:28:52.965135 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8d44b7457-sgd4w"] Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.054893 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-credential-keys\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.054945 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-svc\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.054976 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-combined-ca-bundle\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.054995 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-config\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.055029 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tg4p\" (UniqueName: \"kubernetes.io/projected/3b0ab955-5268-4525-bc85-9723cd1c78e0-kube-api-access-5tg4p\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.055049 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-sb\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.055067 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-fernet-keys\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.055090 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf6dx\" (UniqueName: \"kubernetes.io/projected/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-kube-api-access-cf6dx\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.055111 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-config-data\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.055129 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-nb\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.055155 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-scripts\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.055168 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-swift-storage-0\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.090792 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.092883 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.102401 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.102639 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.125045 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156462 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-combined-ca-bundle\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156515 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-config\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156562 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tg4p\" (UniqueName: \"kubernetes.io/projected/3b0ab955-5268-4525-bc85-9723cd1c78e0-kube-api-access-5tg4p\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156590 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-sb\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156622 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-fernet-keys\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156653 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf6dx\" (UniqueName: \"kubernetes.io/projected/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-kube-api-access-cf6dx\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156678 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-config-data\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156713 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-nb\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156738 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-scripts\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156756 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-swift-storage-0\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156906 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-credential-keys\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.156925 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-svc\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.158717 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-config\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.162962 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-sb\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.163599 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-nb\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.163838 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-swift-storage-0\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.164439 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-svc\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.167975 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-credential-keys\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.168737 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-combined-ca-bundle\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.169544 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-fernet-keys\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.173628 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-config-data\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.182241 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-scripts\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.186017 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tg4p\" (UniqueName: \"kubernetes.io/projected/3b0ab955-5268-4525-bc85-9723cd1c78e0-kube-api-access-5tg4p\") pod \"keystone-bootstrap-b47pf\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.189644 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf6dx\" (UniqueName: \"kubernetes.io/projected/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-kube-api-access-cf6dx\") pod \"dnsmasq-dns-8d44b7457-sgd4w\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.200290 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.222726 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8d44b7457-sgd4w"] Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.223459 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.229093 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zt6f9"] Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.230221 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.242645 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.243350 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.243654 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bqwbw" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.247719 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zt6f9"] Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.258393 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-log-httpd\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.258437 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-scripts\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.258503 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrldk\" (UniqueName: \"kubernetes.io/projected/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-kube-api-access-lrldk\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.258524 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.258576 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-config-data\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.258593 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-run-httpd\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.258635 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.262999 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-k2t47"] Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.264485 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.281054 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-k2t47"] Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.359975 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360050 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-swift-storage-0\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360084 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b7dc36-3f82-4650-b926-03de75dda7e4-logs\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360154 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-config-data\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360180 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-scripts\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360216 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-run-httpd\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360262 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-combined-ca-bundle\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360301 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-nb\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360322 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360339 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-svc\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360380 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-config-data\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.360402 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db9tz\" (UniqueName: \"kubernetes.io/projected/f0b7dc36-3f82-4650-b926-03de75dda7e4-kube-api-access-db9tz\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.361075 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krdjq\" (UniqueName: \"kubernetes.io/projected/65670f97-1301-4482-bb5d-fc4464218f24-kube-api-access-krdjq\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.361108 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-sb\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.361132 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-log-httpd\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.361685 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-scripts\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.362080 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-config\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.362113 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrldk\" (UniqueName: \"kubernetes.io/projected/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-kube-api-access-lrldk\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.361167 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-run-httpd\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.361364 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-log-httpd\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.366541 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.367087 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-scripts\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.368531 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-config-data\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.369377 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.409794 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrldk\" (UniqueName: \"kubernetes.io/projected/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-kube-api-access-lrldk\") pod \"ceilometer-0\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.412839 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.475131 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-config\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.477025 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-swift-storage-0\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.477122 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b7dc36-3f82-4650-b926-03de75dda7e4-logs\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.477331 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-scripts\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.478885 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-swift-storage-0\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.482502 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-nb\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.482585 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-combined-ca-bundle\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.482673 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-svc\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.482721 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-config-data\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.482762 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db9tz\" (UniqueName: \"kubernetes.io/projected/f0b7dc36-3f82-4650-b926-03de75dda7e4-kube-api-access-db9tz\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.482881 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krdjq\" (UniqueName: \"kubernetes.io/projected/65670f97-1301-4482-bb5d-fc4464218f24-kube-api-access-krdjq\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.482919 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-sb\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.484293 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-sb\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.484419 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b7dc36-3f82-4650-b926-03de75dda7e4-logs\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.487433 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-svc\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.488740 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-config\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.492168 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-nb\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.497233 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-combined-ca-bundle\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.504860 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-config-data\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.510004 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-scripts\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.523222 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db9tz\" (UniqueName: \"kubernetes.io/projected/f0b7dc36-3f82-4650-b926-03de75dda7e4-kube-api-access-db9tz\") pod \"placement-db-sync-zt6f9\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.526385 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krdjq\" (UniqueName: \"kubernetes.io/projected/65670f97-1301-4482-bb5d-fc4464218f24-kube-api-access-krdjq\") pod \"dnsmasq-dns-77dd5cf987-k2t47\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.570452 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zt6f9" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.601368 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.687639 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" event={"ID":"eaba3b8a-2131-463a-9dae-1454979ced8b","Type":"ContainerStarted","Data":"4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6"} Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.687845 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" podUID="eaba3b8a-2131-463a-9dae-1454979ced8b" containerName="dnsmasq-dns" containerID="cri-o://4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6" gracePeriod=10 Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.688012 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.717269 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" podStartSLOduration=2.717247471 podStartE2EDuration="2.717247471s" podCreationTimestamp="2025-10-07 22:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:53.711461807 +0000 UTC m=+1207.514159880" watchObservedRunningTime="2025-10-07 22:28:53.717247471 +0000 UTC m=+1207.519945544" Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.781382 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b47pf"] Oct 07 22:28:53 crc kubenswrapper[4871]: I1007 22:28:53.932279 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8d44b7457-sgd4w"] Oct 07 22:28:53 crc kubenswrapper[4871]: W1007 22:28:53.957439 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62c0dc2b_a324_4feb_9910_ac4fa6de7d69.slice/crio-c10bce500dc76cb643f94384b26d0fcc5464892ae1dd15e22ac3399ef4b96a19 WatchSource:0}: Error finding container c10bce500dc76cb643f94384b26d0fcc5464892ae1dd15e22ac3399ef4b96a19: Status 404 returned error can't find the container with id c10bce500dc76cb643f94384b26d0fcc5464892ae1dd15e22ac3399ef4b96a19 Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.006848 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.012249 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.016235 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.016437 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.016861 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8nrnc" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.043368 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.069359 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:28:54 crc kubenswrapper[4871]: W1007 22:28:54.070286 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2daf4b99_374d_45df_b28c_4f02cfa1dfa8.slice/crio-d4b902047b20ab19d87d251e148df5e75c8fe64b09729d10ef1dd00e9d19234a WatchSource:0}: Error finding container d4b902047b20ab19d87d251e148df5e75c8fe64b09729d10ef1dd00e9d19234a: Status 404 returned error can't find the container with id d4b902047b20ab19d87d251e148df5e75c8fe64b09729d10ef1dd00e9d19234a Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.097653 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.100070 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.102852 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.110563 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.151875 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zt6f9"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.196396 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-logs\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.196464 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-config-data\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.196524 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.196749 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.196870 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.196899 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7zkx\" (UniqueName: \"kubernetes.io/projected/d0f1c976-e97d-40b9-a97f-105920655bda-kube-api-access-r7zkx\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.196941 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-scripts\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.233303 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.299512 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.299598 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.299657 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.299708 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.299820 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6m27\" (UniqueName: \"kubernetes.io/projected/9836b5be-f2ba-41fa-81aa-930b045e81f7-kube-api-access-p6m27\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.299853 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.299914 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.299939 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7zkx\" (UniqueName: \"kubernetes.io/projected/d0f1c976-e97d-40b9-a97f-105920655bda-kube-api-access-r7zkx\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.300004 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-scripts\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.300067 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-logs\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.300099 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-config-data\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.300163 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.300262 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-logs\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.300339 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.300933 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.301214 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.303512 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-logs\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.307778 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.310393 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-k2t47"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.310888 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-scripts\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.311685 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-config-data\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.318863 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7zkx\" (UniqueName: \"kubernetes.io/projected/d0f1c976-e97d-40b9-a97f-105920655bda-kube-api-access-r7zkx\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.394975 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f74a-account-create-6gc9r"] Oct 07 22:28:54 crc kubenswrapper[4871]: E1007 22:28:54.396104 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaba3b8a-2131-463a-9dae-1454979ced8b" containerName="init" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.396131 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaba3b8a-2131-463a-9dae-1454979ced8b" containerName="init" Oct 07 22:28:54 crc kubenswrapper[4871]: E1007 22:28:54.396190 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaba3b8a-2131-463a-9dae-1454979ced8b" containerName="dnsmasq-dns" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.396203 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaba3b8a-2131-463a-9dae-1454979ced8b" containerName="dnsmasq-dns" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.396619 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaba3b8a-2131-463a-9dae-1454979ced8b" containerName="dnsmasq-dns" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.397719 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f74a-account-create-6gc9r" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.403029 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-swift-storage-0\") pod \"eaba3b8a-2131-463a-9dae-1454979ced8b\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.403166 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpfp7\" (UniqueName: \"kubernetes.io/projected/eaba3b8a-2131-463a-9dae-1454979ced8b-kube-api-access-jpfp7\") pod \"eaba3b8a-2131-463a-9dae-1454979ced8b\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.403214 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-config\") pod \"eaba3b8a-2131-463a-9dae-1454979ced8b\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.403368 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-sb\") pod \"eaba3b8a-2131-463a-9dae-1454979ced8b\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.403435 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-svc\") pod \"eaba3b8a-2131-463a-9dae-1454979ced8b\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.403475 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-nb\") pod \"eaba3b8a-2131-463a-9dae-1454979ced8b\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.404010 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.404060 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.404114 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6m27\" (UniqueName: \"kubernetes.io/projected/9836b5be-f2ba-41fa-81aa-930b045e81f7-kube-api-access-p6m27\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.404136 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.404280 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-logs\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.404332 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.404362 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.405643 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.408523 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-logs\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.408522 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.412323 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.416585 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.420924 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f74a-account-create-6gc9r"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.424296 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.427427 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.438546 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaba3b8a-2131-463a-9dae-1454979ced8b-kube-api-access-jpfp7" (OuterVolumeSpecName: "kube-api-access-jpfp7") pod "eaba3b8a-2131-463a-9dae-1454979ced8b" (UID: "eaba3b8a-2131-463a-9dae-1454979ced8b"). InnerVolumeSpecName "kube-api-access-jpfp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.439756 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.444690 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6m27\" (UniqueName: \"kubernetes.io/projected/9836b5be-f2ba-41fa-81aa-930b045e81f7-kube-api-access-p6m27\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.553506 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-1599-account-create-ggqbn"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.554659 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1599-account-create-ggqbn" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.556483 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lztcf\" (UniqueName: \"kubernetes.io/projected/cfe40a70-0dba-4ccd-b540-9afbf5685d86-kube-api-access-lztcf\") pod \"cinder-f74a-account-create-6gc9r\" (UID: \"cfe40a70-0dba-4ccd-b540-9afbf5685d86\") " pod="openstack/cinder-f74a-account-create-6gc9r" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.556882 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpfp7\" (UniqueName: \"kubernetes.io/projected/eaba3b8a-2131-463a-9dae-1454979ced8b-kube-api-access-jpfp7\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.566349 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.570411 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1599-account-create-ggqbn"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.644748 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.646685 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.658480 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-config" (OuterVolumeSpecName: "config") pod "eaba3b8a-2131-463a-9dae-1454979ced8b" (UID: "eaba3b8a-2131-463a-9dae-1454979ced8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.658492 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hffsk\" (UniqueName: \"kubernetes.io/projected/39ee3803-7a8c-4a85-b854-d72b5fe17dbb-kube-api-access-hffsk\") pod \"barbican-1599-account-create-ggqbn\" (UID: \"39ee3803-7a8c-4a85-b854-d72b5fe17dbb\") " pod="openstack/barbican-1599-account-create-ggqbn" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.658558 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lztcf\" (UniqueName: \"kubernetes.io/projected/cfe40a70-0dba-4ccd-b540-9afbf5685d86-kube-api-access-lztcf\") pod \"cinder-f74a-account-create-6gc9r\" (UID: \"cfe40a70-0dba-4ccd-b540-9afbf5685d86\") " pod="openstack/cinder-f74a-account-create-6gc9r" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.658634 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.686459 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eaba3b8a-2131-463a-9dae-1454979ced8b" (UID: "eaba3b8a-2131-463a-9dae-1454979ced8b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.692644 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lztcf\" (UniqueName: \"kubernetes.io/projected/cfe40a70-0dba-4ccd-b540-9afbf5685d86-kube-api-access-lztcf\") pod \"cinder-f74a-account-create-6gc9r\" (UID: \"cfe40a70-0dba-4ccd-b540-9afbf5685d86\") " pod="openstack/cinder-f74a-account-create-6gc9r" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.694306 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eaba3b8a-2131-463a-9dae-1454979ced8b" (UID: "eaba3b8a-2131-463a-9dae-1454979ced8b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.711212 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f74a-account-create-6gc9r" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.734860 4871 generic.go:334] "Generic (PLEG): container finished" podID="62c0dc2b-a324-4feb-9910-ac4fa6de7d69" containerID="d8dc3d5bb2b593f501543a8e3d33141b1a8e18963bf195714cf70915fa8bf3bd" exitCode=0 Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.734934 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" event={"ID":"62c0dc2b-a324-4feb-9910-ac4fa6de7d69","Type":"ContainerDied","Data":"d8dc3d5bb2b593f501543a8e3d33141b1a8e18963bf195714cf70915fa8bf3bd"} Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.734967 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" event={"ID":"62c0dc2b-a324-4feb-9910-ac4fa6de7d69","Type":"ContainerStarted","Data":"c10bce500dc76cb643f94384b26d0fcc5464892ae1dd15e22ac3399ef4b96a19"} Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.736472 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eaba3b8a-2131-463a-9dae-1454979ced8b" (UID: "eaba3b8a-2131-463a-9dae-1454979ced8b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.749095 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zt6f9" event={"ID":"f0b7dc36-3f82-4650-b926-03de75dda7e4","Type":"ContainerStarted","Data":"bf6cb3e378505b49aab3cf0349e14ec411e6b52e50a4f92bf77b3f8fb42bdae5"} Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.759386 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eaba3b8a-2131-463a-9dae-1454979ced8b" (UID: "eaba3b8a-2131-463a-9dae-1454979ced8b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.759501 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-swift-storage-0\") pod \"eaba3b8a-2131-463a-9dae-1454979ced8b\" (UID: \"eaba3b8a-2131-463a-9dae-1454979ced8b\") " Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.759991 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hffsk\" (UniqueName: \"kubernetes.io/projected/39ee3803-7a8c-4a85-b854-d72b5fe17dbb-kube-api-access-hffsk\") pod \"barbican-1599-account-create-ggqbn\" (UID: \"39ee3803-7a8c-4a85-b854-d72b5fe17dbb\") " pod="openstack/barbican-1599-account-create-ggqbn" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.760050 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.760062 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.760071 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:54 crc kubenswrapper[4871]: W1007 22:28:54.760428 4871 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/eaba3b8a-2131-463a-9dae-1454979ced8b/volumes/kubernetes.io~configmap/dns-swift-storage-0 Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.760446 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eaba3b8a-2131-463a-9dae-1454979ced8b" (UID: "eaba3b8a-2131-463a-9dae-1454979ced8b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.776719 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.783994 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hffsk\" (UniqueName: \"kubernetes.io/projected/39ee3803-7a8c-4a85-b854-d72b5fe17dbb-kube-api-access-hffsk\") pod \"barbican-1599-account-create-ggqbn\" (UID: \"39ee3803-7a8c-4a85-b854-d72b5fe17dbb\") " pod="openstack/barbican-1599-account-create-ggqbn" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.789497 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b47pf" event={"ID":"3b0ab955-5268-4525-bc85-9723cd1c78e0","Type":"ContainerStarted","Data":"15a075e9bb05d488c246bfc9ab9f1e3bce81bc15386a51cca52b6ab00f6b66e8"} Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.789559 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b47pf" event={"ID":"3b0ab955-5268-4525-bc85-9723cd1c78e0","Type":"ContainerStarted","Data":"2eb550c70e0a0d5954545ae4423c833d3bf773e3efbb48afedf77a976a873735"} Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.792268 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-719b-account-create-bx5fz"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.793485 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-719b-account-create-bx5fz" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.793690 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" event={"ID":"65670f97-1301-4482-bb5d-fc4464218f24","Type":"ContainerStarted","Data":"746a259f96ec23e61b97a16b0ba85c0ca471d0c11abd75dca9fdd979e5d11b0f"} Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.795942 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.820474 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-719b-account-create-bx5fz"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.821742 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-b47pf" podStartSLOduration=2.821723184 podStartE2EDuration="2.821723184s" podCreationTimestamp="2025-10-07 22:28:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:54.808629277 +0000 UTC m=+1208.611327350" watchObservedRunningTime="2025-10-07 22:28:54.821723184 +0000 UTC m=+1208.624421257" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.825497 4871 generic.go:334] "Generic (PLEG): container finished" podID="eaba3b8a-2131-463a-9dae-1454979ced8b" containerID="4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6" exitCode=0 Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.825566 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" event={"ID":"eaba3b8a-2131-463a-9dae-1454979ced8b","Type":"ContainerDied","Data":"4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6"} Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.825597 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" event={"ID":"eaba3b8a-2131-463a-9dae-1454979ced8b","Type":"ContainerDied","Data":"a0bc0a1485e7fea4de23737f1d61e6300a251e146592e2b44ac64f778bb0be0c"} Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.825626 4871 scope.go:117] "RemoveContainer" containerID="4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.825781 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf6f78f57-x7clx" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.835571 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerStarted","Data":"d4b902047b20ab19d87d251e148df5e75c8fe64b09729d10ef1dd00e9d19234a"} Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.866290 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnz62\" (UniqueName: \"kubernetes.io/projected/0a22909c-2279-4ea0-96b4-9abce934edc3-kube-api-access-rnz62\") pod \"neutron-719b-account-create-bx5fz\" (UID: \"0a22909c-2279-4ea0-96b4-9abce934edc3\") " pod="openstack/neutron-719b-account-create-bx5fz" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.866891 4871 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eaba3b8a-2131-463a-9dae-1454979ced8b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.876878 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf6f78f57-x7clx"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.890724 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf6f78f57-x7clx"] Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.894963 4871 scope.go:117] "RemoveContainer" containerID="115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.956018 4871 scope.go:117] "RemoveContainer" containerID="4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6" Oct 07 22:28:54 crc kubenswrapper[4871]: E1007 22:28:54.961699 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6\": container with ID starting with 4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6 not found: ID does not exist" containerID="4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.961751 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6"} err="failed to get container status \"4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6\": rpc error: code = NotFound desc = could not find container \"4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6\": container with ID starting with 4d1f587e9e90426581bce68c3f6525596aeb0ea5d672c886481a2b662eb005a6 not found: ID does not exist" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.961782 4871 scope.go:117] "RemoveContainer" containerID="115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901" Oct 07 22:28:54 crc kubenswrapper[4871]: E1007 22:28:54.962344 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901\": container with ID starting with 115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901 not found: ID does not exist" containerID="115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.962391 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901"} err="failed to get container status \"115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901\": rpc error: code = NotFound desc = could not find container \"115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901\": container with ID starting with 115aa30633474f566f83e1f18cb54b533bbbef728411a212bbd9fc1c697bf901 not found: ID does not exist" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.968217 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnz62\" (UniqueName: \"kubernetes.io/projected/0a22909c-2279-4ea0-96b4-9abce934edc3-kube-api-access-rnz62\") pod \"neutron-719b-account-create-bx5fz\" (UID: \"0a22909c-2279-4ea0-96b4-9abce934edc3\") " pod="openstack/neutron-719b-account-create-bx5fz" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.989743 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnz62\" (UniqueName: \"kubernetes.io/projected/0a22909c-2279-4ea0-96b4-9abce934edc3-kube-api-access-rnz62\") pod \"neutron-719b-account-create-bx5fz\" (UID: \"0a22909c-2279-4ea0-96b4-9abce934edc3\") " pod="openstack/neutron-719b-account-create-bx5fz" Oct 07 22:28:54 crc kubenswrapper[4871]: I1007 22:28:54.996413 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaba3b8a-2131-463a-9dae-1454979ced8b" path="/var/lib/kubelet/pods/eaba3b8a-2131-463a-9dae-1454979ced8b/volumes" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.034531 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1599-account-create-ggqbn" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.117970 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-719b-account-create-bx5fz" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.204355 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.281557 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-nb\") pod \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.281929 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-svc\") pod \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.281950 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-sb\") pod \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.282001 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-config\") pod \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.282133 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf6dx\" (UniqueName: \"kubernetes.io/projected/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-kube-api-access-cf6dx\") pod \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.282182 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-swift-storage-0\") pod \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\" (UID: \"62c0dc2b-a324-4feb-9910-ac4fa6de7d69\") " Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.287469 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-kube-api-access-cf6dx" (OuterVolumeSpecName: "kube-api-access-cf6dx") pod "62c0dc2b-a324-4feb-9910-ac4fa6de7d69" (UID: "62c0dc2b-a324-4feb-9910-ac4fa6de7d69"). InnerVolumeSpecName "kube-api-access-cf6dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.308340 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "62c0dc2b-a324-4feb-9910-ac4fa6de7d69" (UID: "62c0dc2b-a324-4feb-9910-ac4fa6de7d69"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.317286 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "62c0dc2b-a324-4feb-9910-ac4fa6de7d69" (UID: "62c0dc2b-a324-4feb-9910-ac4fa6de7d69"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.322593 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "62c0dc2b-a324-4feb-9910-ac4fa6de7d69" (UID: "62c0dc2b-a324-4feb-9910-ac4fa6de7d69"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.323671 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-config" (OuterVolumeSpecName: "config") pod "62c0dc2b-a324-4feb-9910-ac4fa6de7d69" (UID: "62c0dc2b-a324-4feb-9910-ac4fa6de7d69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.341280 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "62c0dc2b-a324-4feb-9910-ac4fa6de7d69" (UID: "62c0dc2b-a324-4feb-9910-ac4fa6de7d69"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.384677 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.384719 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.384732 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.384746 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.384760 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf6dx\" (UniqueName: \"kubernetes.io/projected/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-kube-api-access-cf6dx\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.384774 4871 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/62c0dc2b-a324-4feb-9910-ac4fa6de7d69-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.413642 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f74a-account-create-6gc9r"] Oct 07 22:28:55 crc kubenswrapper[4871]: W1007 22:28:55.416921 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfe40a70_0dba_4ccd_b540_9afbf5685d86.slice/crio-7ae9468cd50c9d125ac67cbeaaf67d0645f4dd76ab468d6e80b9c97cc1caa08a WatchSource:0}: Error finding container 7ae9468cd50c9d125ac67cbeaaf67d0645f4dd76ab468d6e80b9c97cc1caa08a: Status 404 returned error can't find the container with id 7ae9468cd50c9d125ac67cbeaaf67d0645f4dd76ab468d6e80b9c97cc1caa08a Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.526298 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.655536 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:28:55 crc kubenswrapper[4871]: W1007 22:28:55.665501 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9836b5be_f2ba_41fa_81aa_930b045e81f7.slice/crio-6cd6ec4371fdb6a8137d569600bc1f9f870415ca8400dfc86f61b1d2cb8d456c WatchSource:0}: Error finding container 6cd6ec4371fdb6a8137d569600bc1f9f870415ca8400dfc86f61b1d2cb8d456c: Status 404 returned error can't find the container with id 6cd6ec4371fdb6a8137d569600bc1f9f870415ca8400dfc86f61b1d2cb8d456c Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.693529 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1599-account-create-ggqbn"] Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.788178 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-719b-account-create-bx5fz"] Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.851604 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1599-account-create-ggqbn" event={"ID":"39ee3803-7a8c-4a85-b854-d72b5fe17dbb","Type":"ContainerStarted","Data":"67e46cb719e653a418f799bf2e5087cc54d0cd7fb49b8824f0d639478ad2417c"} Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.853097 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0f1c976-e97d-40b9-a97f-105920655bda","Type":"ContainerStarted","Data":"dffc963a964e2b5142825ca6444b3926b2217bc59145e8a3411f3b9f5a13a93c"} Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.861478 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" event={"ID":"62c0dc2b-a324-4feb-9910-ac4fa6de7d69","Type":"ContainerDied","Data":"c10bce500dc76cb643f94384b26d0fcc5464892ae1dd15e22ac3399ef4b96a19"} Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.864469 4871 scope.go:117] "RemoveContainer" containerID="d8dc3d5bb2b593f501543a8e3d33141b1a8e18963bf195714cf70915fa8bf3bd" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.864817 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8d44b7457-sgd4w" Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.875946 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-719b-account-create-bx5fz" event={"ID":"0a22909c-2279-4ea0-96b4-9abce934edc3","Type":"ContainerStarted","Data":"cd52118339466fc4582b50c1ac94e4f6a0e0ba14c95e1eadd7ba160e29751549"} Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.880418 4871 generic.go:334] "Generic (PLEG): container finished" podID="65670f97-1301-4482-bb5d-fc4464218f24" containerID="7752dee107e8a50cc0896cfb6745ee58b4e5272b92b6a05daf6722296ef556c7" exitCode=0 Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.881018 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" event={"ID":"65670f97-1301-4482-bb5d-fc4464218f24","Type":"ContainerDied","Data":"7752dee107e8a50cc0896cfb6745ee58b4e5272b92b6a05daf6722296ef556c7"} Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.891556 4871 generic.go:334] "Generic (PLEG): container finished" podID="cfe40a70-0dba-4ccd-b540-9afbf5685d86" containerID="db14634af4bd8ffd95be4d16f4dbd649bbf43ca5ba952ff1a56f66dcbab97868" exitCode=0 Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.891669 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f74a-account-create-6gc9r" event={"ID":"cfe40a70-0dba-4ccd-b540-9afbf5685d86","Type":"ContainerDied","Data":"db14634af4bd8ffd95be4d16f4dbd649bbf43ca5ba952ff1a56f66dcbab97868"} Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.891704 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f74a-account-create-6gc9r" event={"ID":"cfe40a70-0dba-4ccd-b540-9afbf5685d86","Type":"ContainerStarted","Data":"7ae9468cd50c9d125ac67cbeaaf67d0645f4dd76ab468d6e80b9c97cc1caa08a"} Oct 07 22:28:55 crc kubenswrapper[4871]: I1007 22:28:55.978162 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9836b5be-f2ba-41fa-81aa-930b045e81f7","Type":"ContainerStarted","Data":"6cd6ec4371fdb6a8137d569600bc1f9f870415ca8400dfc86f61b1d2cb8d456c"} Oct 07 22:28:56 crc kubenswrapper[4871]: I1007 22:28:56.018562 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8d44b7457-sgd4w"] Oct 07 22:28:56 crc kubenswrapper[4871]: I1007 22:28:56.040518 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8d44b7457-sgd4w"] Oct 07 22:28:56 crc kubenswrapper[4871]: I1007 22:28:56.913134 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:28:56 crc kubenswrapper[4871]: I1007 22:28:56.943048 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.051844 4871 generic.go:334] "Generic (PLEG): container finished" podID="39ee3803-7a8c-4a85-b854-d72b5fe17dbb" containerID="a4fb5e314a418ad3f5df08f22eefe7f38b31a60269157a06a8f3094f45ee90eb" exitCode=0 Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.059634 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62c0dc2b-a324-4feb-9910-ac4fa6de7d69" path="/var/lib/kubelet/pods/62c0dc2b-a324-4feb-9910-ac4fa6de7d69/volumes" Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.060804 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9836b5be-f2ba-41fa-81aa-930b045e81f7","Type":"ContainerStarted","Data":"8ed9776a16b485c2c2642bcabf5d61e86921b8b52f37c2610b01e764373ee95c"} Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.060843 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1599-account-create-ggqbn" event={"ID":"39ee3803-7a8c-4a85-b854-d72b5fe17dbb","Type":"ContainerDied","Data":"a4fb5e314a418ad3f5df08f22eefe7f38b31a60269157a06a8f3094f45ee90eb"} Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.060870 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.072344 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0f1c976-e97d-40b9-a97f-105920655bda","Type":"ContainerStarted","Data":"650b1a6f51292f66150c255a0c37fc55a4deb4ccee8e73d591fe49e09f2a0754"} Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.084460 4871 generic.go:334] "Generic (PLEG): container finished" podID="0a22909c-2279-4ea0-96b4-9abce934edc3" containerID="0c8b1c9fd16dade7ae8c609c61cead66d6386df85d637f9397df20edbb885c4a" exitCode=0 Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.084548 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-719b-account-create-bx5fz" event={"ID":"0a22909c-2279-4ea0-96b4-9abce934edc3","Type":"ContainerDied","Data":"0c8b1c9fd16dade7ae8c609c61cead66d6386df85d637f9397df20edbb885c4a"} Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.098099 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" event={"ID":"65670f97-1301-4482-bb5d-fc4464218f24","Type":"ContainerStarted","Data":"d0390f1f00eaa4fb6e4dc40b42d55096049958ee99bfee6fbcb90c72fc6aac1a"} Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.098617 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.194983 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" podStartSLOduration=4.194955535 podStartE2EDuration="4.194955535s" podCreationTimestamp="2025-10-07 22:28:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:57.192641133 +0000 UTC m=+1210.995339196" watchObservedRunningTime="2025-10-07 22:28:57.194955535 +0000 UTC m=+1210.997653608" Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.497156 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f74a-account-create-6gc9r" Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.569889 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lztcf\" (UniqueName: \"kubernetes.io/projected/cfe40a70-0dba-4ccd-b540-9afbf5685d86-kube-api-access-lztcf\") pod \"cfe40a70-0dba-4ccd-b540-9afbf5685d86\" (UID: \"cfe40a70-0dba-4ccd-b540-9afbf5685d86\") " Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.578448 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe40a70-0dba-4ccd-b540-9afbf5685d86-kube-api-access-lztcf" (OuterVolumeSpecName: "kube-api-access-lztcf") pod "cfe40a70-0dba-4ccd-b540-9afbf5685d86" (UID: "cfe40a70-0dba-4ccd-b540-9afbf5685d86"). InnerVolumeSpecName "kube-api-access-lztcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:28:57 crc kubenswrapper[4871]: I1007 22:28:57.673250 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lztcf\" (UniqueName: \"kubernetes.io/projected/cfe40a70-0dba-4ccd-b540-9afbf5685d86-kube-api-access-lztcf\") on node \"crc\" DevicePath \"\"" Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.114194 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0f1c976-e97d-40b9-a97f-105920655bda","Type":"ContainerStarted","Data":"cf553428752c5df543a55de342dcca65031384272035f4e1dedff261e60e8752"} Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.114320 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d0f1c976-e97d-40b9-a97f-105920655bda" containerName="glance-log" containerID="cri-o://650b1a6f51292f66150c255a0c37fc55a4deb4ccee8e73d591fe49e09f2a0754" gracePeriod=30 Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.114670 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d0f1c976-e97d-40b9-a97f-105920655bda" containerName="glance-httpd" containerID="cri-o://cf553428752c5df543a55de342dcca65031384272035f4e1dedff261e60e8752" gracePeriod=30 Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.136112 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f74a-account-create-6gc9r" event={"ID":"cfe40a70-0dba-4ccd-b540-9afbf5685d86","Type":"ContainerDied","Data":"7ae9468cd50c9d125ac67cbeaaf67d0645f4dd76ab468d6e80b9c97cc1caa08a"} Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.136155 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ae9468cd50c9d125ac67cbeaaf67d0645f4dd76ab468d6e80b9c97cc1caa08a" Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.136308 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f74a-account-create-6gc9r" Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.148545 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9836b5be-f2ba-41fa-81aa-930b045e81f7","Type":"ContainerStarted","Data":"6df113a7ac9f9d5f536d35f13f7fad51d2a38b5a2ccec4a5d4e7b944182cadb4"} Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.149446 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerName="glance-log" containerID="cri-o://8ed9776a16b485c2c2642bcabf5d61e86921b8b52f37c2610b01e764373ee95c" gracePeriod=30 Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.149590 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerName="glance-httpd" containerID="cri-o://6df113a7ac9f9d5f536d35f13f7fad51d2a38b5a2ccec4a5d4e7b944182cadb4" gracePeriod=30 Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.185079 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.185060876 podStartE2EDuration="6.185060876s" podCreationTimestamp="2025-10-07 22:28:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:58.155989495 +0000 UTC m=+1211.958687558" watchObservedRunningTime="2025-10-07 22:28:58.185060876 +0000 UTC m=+1211.987758949" Oct 07 22:28:58 crc kubenswrapper[4871]: I1007 22:28:58.187301 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.187294275 podStartE2EDuration="5.187294275s" podCreationTimestamp="2025-10-07 22:28:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:28:58.186652168 +0000 UTC m=+1211.989350241" watchObservedRunningTime="2025-10-07 22:28:58.187294275 +0000 UTC m=+1211.989992348" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.160309 4871 generic.go:334] "Generic (PLEG): container finished" podID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerID="6df113a7ac9f9d5f536d35f13f7fad51d2a38b5a2ccec4a5d4e7b944182cadb4" exitCode=0 Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.160671 4871 generic.go:334] "Generic (PLEG): container finished" podID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerID="8ed9776a16b485c2c2642bcabf5d61e86921b8b52f37c2610b01e764373ee95c" exitCode=143 Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.160448 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9836b5be-f2ba-41fa-81aa-930b045e81f7","Type":"ContainerDied","Data":"6df113a7ac9f9d5f536d35f13f7fad51d2a38b5a2ccec4a5d4e7b944182cadb4"} Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.160749 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9836b5be-f2ba-41fa-81aa-930b045e81f7","Type":"ContainerDied","Data":"8ed9776a16b485c2c2642bcabf5d61e86921b8b52f37c2610b01e764373ee95c"} Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.164467 4871 generic.go:334] "Generic (PLEG): container finished" podID="d0f1c976-e97d-40b9-a97f-105920655bda" containerID="cf553428752c5df543a55de342dcca65031384272035f4e1dedff261e60e8752" exitCode=0 Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.164520 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0f1c976-e97d-40b9-a97f-105920655bda","Type":"ContainerDied","Data":"cf553428752c5df543a55de342dcca65031384272035f4e1dedff261e60e8752"} Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.164528 4871 generic.go:334] "Generic (PLEG): container finished" podID="d0f1c976-e97d-40b9-a97f-105920655bda" containerID="650b1a6f51292f66150c255a0c37fc55a4deb4ccee8e73d591fe49e09f2a0754" exitCode=143 Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.164549 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0f1c976-e97d-40b9-a97f-105920655bda","Type":"ContainerDied","Data":"650b1a6f51292f66150c255a0c37fc55a4deb4ccee8e73d591fe49e09f2a0754"} Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.167614 4871 generic.go:334] "Generic (PLEG): container finished" podID="3b0ab955-5268-4525-bc85-9723cd1c78e0" containerID="15a075e9bb05d488c246bfc9ab9f1e3bce81bc15386a51cca52b6ab00f6b66e8" exitCode=0 Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.168478 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b47pf" event={"ID":"3b0ab955-5268-4525-bc85-9723cd1c78e0","Type":"ContainerDied","Data":"15a075e9bb05d488c246bfc9ab9f1e3bce81bc15386a51cca52b6ab00f6b66e8"} Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.664385 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7sq86"] Oct 07 22:28:59 crc kubenswrapper[4871]: E1007 22:28:59.664949 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c0dc2b-a324-4feb-9910-ac4fa6de7d69" containerName="init" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.664970 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c0dc2b-a324-4feb-9910-ac4fa6de7d69" containerName="init" Oct 07 22:28:59 crc kubenswrapper[4871]: E1007 22:28:59.665015 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe40a70-0dba-4ccd-b540-9afbf5685d86" containerName="mariadb-account-create" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.665024 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe40a70-0dba-4ccd-b540-9afbf5685d86" containerName="mariadb-account-create" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.665247 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe40a70-0dba-4ccd-b540-9afbf5685d86" containerName="mariadb-account-create" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.665277 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c0dc2b-a324-4feb-9910-ac4fa6de7d69" containerName="init" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.666032 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.674027 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.674236 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-lhvh2" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.674610 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7sq86"] Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.680191 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.722503 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-combined-ca-bundle\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.722598 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-etc-machine-id\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.722670 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hps9v\" (UniqueName: \"kubernetes.io/projected/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-kube-api-access-hps9v\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.722719 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-scripts\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.723263 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-config-data\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.723458 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-db-sync-config-data\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.826088 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-config-data\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.826198 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-db-sync-config-data\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.826290 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-combined-ca-bundle\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.826331 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-etc-machine-id\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.826364 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hps9v\" (UniqueName: \"kubernetes.io/projected/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-kube-api-access-hps9v\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.826395 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-scripts\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.827946 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-etc-machine-id\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.834249 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-db-sync-config-data\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.834893 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-scripts\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.835144 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-combined-ca-bundle\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.835534 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-config-data\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:28:59 crc kubenswrapper[4871]: I1007 22:28:59.845972 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hps9v\" (UniqueName: \"kubernetes.io/projected/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-kube-api-access-hps9v\") pod \"cinder-db-sync-7sq86\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " pod="openstack/cinder-db-sync-7sq86" Oct 07 22:29:00 crc kubenswrapper[4871]: I1007 22:29:00.039069 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7sq86" Oct 07 22:29:00 crc kubenswrapper[4871]: I1007 22:29:00.873230 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-719b-account-create-bx5fz" Oct 07 22:29:00 crc kubenswrapper[4871]: I1007 22:29:00.880195 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1599-account-create-ggqbn" Oct 07 22:29:00 crc kubenswrapper[4871]: I1007 22:29:00.950348 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hffsk\" (UniqueName: \"kubernetes.io/projected/39ee3803-7a8c-4a85-b854-d72b5fe17dbb-kube-api-access-hffsk\") pod \"39ee3803-7a8c-4a85-b854-d72b5fe17dbb\" (UID: \"39ee3803-7a8c-4a85-b854-d72b5fe17dbb\") " Oct 07 22:29:00 crc kubenswrapper[4871]: I1007 22:29:00.950572 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnz62\" (UniqueName: \"kubernetes.io/projected/0a22909c-2279-4ea0-96b4-9abce934edc3-kube-api-access-rnz62\") pod \"0a22909c-2279-4ea0-96b4-9abce934edc3\" (UID: \"0a22909c-2279-4ea0-96b4-9abce934edc3\") " Oct 07 22:29:00 crc kubenswrapper[4871]: I1007 22:29:00.968029 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39ee3803-7a8c-4a85-b854-d72b5fe17dbb-kube-api-access-hffsk" (OuterVolumeSpecName: "kube-api-access-hffsk") pod "39ee3803-7a8c-4a85-b854-d72b5fe17dbb" (UID: "39ee3803-7a8c-4a85-b854-d72b5fe17dbb"). InnerVolumeSpecName "kube-api-access-hffsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:00 crc kubenswrapper[4871]: I1007 22:29:00.971054 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a22909c-2279-4ea0-96b4-9abce934edc3-kube-api-access-rnz62" (OuterVolumeSpecName: "kube-api-access-rnz62") pod "0a22909c-2279-4ea0-96b4-9abce934edc3" (UID: "0a22909c-2279-4ea0-96b4-9abce934edc3"). InnerVolumeSpecName "kube-api-access-rnz62". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.054047 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hffsk\" (UniqueName: \"kubernetes.io/projected/39ee3803-7a8c-4a85-b854-d72b5fe17dbb-kube-api-access-hffsk\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.054071 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnz62\" (UniqueName: \"kubernetes.io/projected/0a22909c-2279-4ea0-96b4-9abce934edc3-kube-api-access-rnz62\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.065889 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.158217 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tg4p\" (UniqueName: \"kubernetes.io/projected/3b0ab955-5268-4525-bc85-9723cd1c78e0-kube-api-access-5tg4p\") pod \"3b0ab955-5268-4525-bc85-9723cd1c78e0\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.158317 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-credential-keys\") pod \"3b0ab955-5268-4525-bc85-9723cd1c78e0\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.158374 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-fernet-keys\") pod \"3b0ab955-5268-4525-bc85-9723cd1c78e0\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.158416 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-scripts\") pod \"3b0ab955-5268-4525-bc85-9723cd1c78e0\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.158531 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-combined-ca-bundle\") pod \"3b0ab955-5268-4525-bc85-9723cd1c78e0\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.158551 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-config-data\") pod \"3b0ab955-5268-4525-bc85-9723cd1c78e0\" (UID: \"3b0ab955-5268-4525-bc85-9723cd1c78e0\") " Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.163609 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3b0ab955-5268-4525-bc85-9723cd1c78e0" (UID: "3b0ab955-5268-4525-bc85-9723cd1c78e0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.178778 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3b0ab955-5268-4525-bc85-9723cd1c78e0" (UID: "3b0ab955-5268-4525-bc85-9723cd1c78e0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.187061 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0ab955-5268-4525-bc85-9723cd1c78e0-kube-api-access-5tg4p" (OuterVolumeSpecName: "kube-api-access-5tg4p") pod "3b0ab955-5268-4525-bc85-9723cd1c78e0" (UID: "3b0ab955-5268-4525-bc85-9723cd1c78e0"). InnerVolumeSpecName "kube-api-access-5tg4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.194735 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-scripts" (OuterVolumeSpecName: "scripts") pod "3b0ab955-5268-4525-bc85-9723cd1c78e0" (UID: "3b0ab955-5268-4525-bc85-9723cd1c78e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.204636 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-config-data" (OuterVolumeSpecName: "config-data") pod "3b0ab955-5268-4525-bc85-9723cd1c78e0" (UID: "3b0ab955-5268-4525-bc85-9723cd1c78e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.206133 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1599-account-create-ggqbn" event={"ID":"39ee3803-7a8c-4a85-b854-d72b5fe17dbb","Type":"ContainerDied","Data":"67e46cb719e653a418f799bf2e5087cc54d0cd7fb49b8824f0d639478ad2417c"} Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.206185 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67e46cb719e653a418f799bf2e5087cc54d0cd7fb49b8824f0d639478ad2417c" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.206281 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1599-account-create-ggqbn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.227010 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b0ab955-5268-4525-bc85-9723cd1c78e0" (UID: "3b0ab955-5268-4525-bc85-9723cd1c78e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.231250 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-719b-account-create-bx5fz" event={"ID":"0a22909c-2279-4ea0-96b4-9abce934edc3","Type":"ContainerDied","Data":"cd52118339466fc4582b50c1ac94e4f6a0e0ba14c95e1eadd7ba160e29751549"} Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.231298 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd52118339466fc4582b50c1ac94e4f6a0e0ba14c95e1eadd7ba160e29751549" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.231417 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-719b-account-create-bx5fz" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.254441 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b47pf" event={"ID":"3b0ab955-5268-4525-bc85-9723cd1c78e0","Type":"ContainerDied","Data":"2eb550c70e0a0d5954545ae4423c833d3bf773e3efbb48afedf77a976a873735"} Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.254488 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2eb550c70e0a0d5954545ae4423c833d3bf773e3efbb48afedf77a976a873735" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.254552 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b47pf" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.261684 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.261719 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.261730 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.261740 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tg4p\" (UniqueName: \"kubernetes.io/projected/3b0ab955-5268-4525-bc85-9723cd1c78e0-kube-api-access-5tg4p\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.261749 4871 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.261758 4871 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b0ab955-5268-4525-bc85-9723cd1c78e0-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.403878 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-b47pf"] Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.422520 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-b47pf"] Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.497301 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-7zfdn"] Oct 07 22:29:01 crc kubenswrapper[4871]: E1007 22:29:01.497775 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a22909c-2279-4ea0-96b4-9abce934edc3" containerName="mariadb-account-create" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.505875 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a22909c-2279-4ea0-96b4-9abce934edc3" containerName="mariadb-account-create" Oct 07 22:29:01 crc kubenswrapper[4871]: E1007 22:29:01.505952 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ee3803-7a8c-4a85-b854-d72b5fe17dbb" containerName="mariadb-account-create" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.505967 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ee3803-7a8c-4a85-b854-d72b5fe17dbb" containerName="mariadb-account-create" Oct 07 22:29:01 crc kubenswrapper[4871]: E1007 22:29:01.506040 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0ab955-5268-4525-bc85-9723cd1c78e0" containerName="keystone-bootstrap" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.506050 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0ab955-5268-4525-bc85-9723cd1c78e0" containerName="keystone-bootstrap" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.506450 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ee3803-7a8c-4a85-b854-d72b5fe17dbb" containerName="mariadb-account-create" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.506471 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b0ab955-5268-4525-bc85-9723cd1c78e0" containerName="keystone-bootstrap" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.506482 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a22909c-2279-4ea0-96b4-9abce934edc3" containerName="mariadb-account-create" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.511426 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.515463 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4kmth" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.515610 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.515765 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.541889 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.565478 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7zfdn"] Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.572077 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwzcp\" (UniqueName: \"kubernetes.io/projected/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-kube-api-access-kwzcp\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.572126 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-combined-ca-bundle\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.572154 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-config-data\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.572184 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-scripts\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.572199 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-credential-keys\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.572269 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-fernet-keys\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.674125 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-credential-keys\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.674207 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-scripts\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.674319 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-fernet-keys\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.674442 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwzcp\" (UniqueName: \"kubernetes.io/projected/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-kube-api-access-kwzcp\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.674478 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-combined-ca-bundle\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.674513 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-config-data\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.680713 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-scripts\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.681000 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-credential-keys\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.681239 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-fernet-keys\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.684777 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-config-data\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.686229 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-combined-ca-bundle\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.691537 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwzcp\" (UniqueName: \"kubernetes.io/projected/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-kube-api-access-kwzcp\") pod \"keystone-bootstrap-7zfdn\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:01 crc kubenswrapper[4871]: I1007 22:29:01.870959 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:02.996702 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b0ab955-5268-4525-bc85-9723cd1c78e0" path="/var/lib/kubelet/pods/3b0ab955-5268-4525-bc85-9723cd1c78e0/volumes" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.289513 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9836b5be-f2ba-41fa-81aa-930b045e81f7","Type":"ContainerDied","Data":"6cd6ec4371fdb6a8137d569600bc1f9f870415ca8400dfc86f61b1d2cb8d456c"} Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.289964 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cd6ec4371fdb6a8137d569600bc1f9f870415ca8400dfc86f61b1d2cb8d456c" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.293172 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0f1c976-e97d-40b9-a97f-105920655bda","Type":"ContainerDied","Data":"dffc963a964e2b5142825ca6444b3926b2217bc59145e8a3411f3b9f5a13a93c"} Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.293227 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dffc963a964e2b5142825ca6444b3926b2217bc59145e8a3411f3b9f5a13a93c" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.355469 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.364333 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415137 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-config-data\") pod \"d0f1c976-e97d-40b9-a97f-105920655bda\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415210 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-scripts\") pod \"9836b5be-f2ba-41fa-81aa-930b045e81f7\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415251 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6m27\" (UniqueName: \"kubernetes.io/projected/9836b5be-f2ba-41fa-81aa-930b045e81f7-kube-api-access-p6m27\") pod \"9836b5be-f2ba-41fa-81aa-930b045e81f7\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415302 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-scripts\") pod \"d0f1c976-e97d-40b9-a97f-105920655bda\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415327 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-combined-ca-bundle\") pod \"d0f1c976-e97d-40b9-a97f-105920655bda\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415384 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7zkx\" (UniqueName: \"kubernetes.io/projected/d0f1c976-e97d-40b9-a97f-105920655bda-kube-api-access-r7zkx\") pod \"d0f1c976-e97d-40b9-a97f-105920655bda\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415419 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"d0f1c976-e97d-40b9-a97f-105920655bda\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415443 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-config-data\") pod \"9836b5be-f2ba-41fa-81aa-930b045e81f7\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415470 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-logs\") pod \"d0f1c976-e97d-40b9-a97f-105920655bda\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415501 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-httpd-run\") pod \"9836b5be-f2ba-41fa-81aa-930b045e81f7\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415526 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-httpd-run\") pod \"d0f1c976-e97d-40b9-a97f-105920655bda\" (UID: \"d0f1c976-e97d-40b9-a97f-105920655bda\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415608 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-combined-ca-bundle\") pod \"9836b5be-f2ba-41fa-81aa-930b045e81f7\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415656 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"9836b5be-f2ba-41fa-81aa-930b045e81f7\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.415683 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-logs\") pod \"9836b5be-f2ba-41fa-81aa-930b045e81f7\" (UID: \"9836b5be-f2ba-41fa-81aa-930b045e81f7\") " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.416694 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-logs" (OuterVolumeSpecName: "logs") pod "9836b5be-f2ba-41fa-81aa-930b045e81f7" (UID: "9836b5be-f2ba-41fa-81aa-930b045e81f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.419951 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d0f1c976-e97d-40b9-a97f-105920655bda" (UID: "d0f1c976-e97d-40b9-a97f-105920655bda"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.420229 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-logs" (OuterVolumeSpecName: "logs") pod "d0f1c976-e97d-40b9-a97f-105920655bda" (UID: "d0f1c976-e97d-40b9-a97f-105920655bda"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.420476 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9836b5be-f2ba-41fa-81aa-930b045e81f7" (UID: "9836b5be-f2ba-41fa-81aa-930b045e81f7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.423738 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "d0f1c976-e97d-40b9-a97f-105920655bda" (UID: "d0f1c976-e97d-40b9-a97f-105920655bda"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.446926 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9836b5be-f2ba-41fa-81aa-930b045e81f7-kube-api-access-p6m27" (OuterVolumeSpecName: "kube-api-access-p6m27") pod "9836b5be-f2ba-41fa-81aa-930b045e81f7" (UID: "9836b5be-f2ba-41fa-81aa-930b045e81f7"). InnerVolumeSpecName "kube-api-access-p6m27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.451245 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-scripts" (OuterVolumeSpecName: "scripts") pod "9836b5be-f2ba-41fa-81aa-930b045e81f7" (UID: "9836b5be-f2ba-41fa-81aa-930b045e81f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.455521 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-scripts" (OuterVolumeSpecName: "scripts") pod "d0f1c976-e97d-40b9-a97f-105920655bda" (UID: "d0f1c976-e97d-40b9-a97f-105920655bda"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.468185 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "9836b5be-f2ba-41fa-81aa-930b045e81f7" (UID: "9836b5be-f2ba-41fa-81aa-930b045e81f7"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.482134 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0f1c976-e97d-40b9-a97f-105920655bda-kube-api-access-r7zkx" (OuterVolumeSpecName: "kube-api-access-r7zkx") pod "d0f1c976-e97d-40b9-a97f-105920655bda" (UID: "d0f1c976-e97d-40b9-a97f-105920655bda"). InnerVolumeSpecName "kube-api-access-r7zkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.517326 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.517611 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.517735 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.517834 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6m27\" (UniqueName: \"kubernetes.io/projected/9836b5be-f2ba-41fa-81aa-930b045e81f7-kube-api-access-p6m27\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.517908 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.517996 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7zkx\" (UniqueName: \"kubernetes.io/projected/d0f1c976-e97d-40b9-a97f-105920655bda-kube-api-access-r7zkx\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.518073 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.518134 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.518184 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9836b5be-f2ba-41fa-81aa-930b045e81f7-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.518260 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0f1c976-e97d-40b9-a97f-105920655bda-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.520875 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0f1c976-e97d-40b9-a97f-105920655bda" (UID: "d0f1c976-e97d-40b9-a97f-105920655bda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.532426 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9836b5be-f2ba-41fa-81aa-930b045e81f7" (UID: "9836b5be-f2ba-41fa-81aa-930b045e81f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.540325 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.574181 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.578329 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-config-data" (OuterVolumeSpecName: "config-data") pod "9836b5be-f2ba-41fa-81aa-930b045e81f7" (UID: "9836b5be-f2ba-41fa-81aa-930b045e81f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.600726 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-config-data" (OuterVolumeSpecName: "config-data") pod "d0f1c976-e97d-40b9-a97f-105920655bda" (UID: "d0f1c976-e97d-40b9-a97f-105920655bda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.603527 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.620957 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.621006 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.621023 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.621037 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9836b5be-f2ba-41fa-81aa-930b045e81f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.621050 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.621062 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f1c976-e97d-40b9-a97f-105920655bda-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.673557 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-f26k9"] Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.673867 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" podUID="63e41063-fa11-4824-87df-49c82c6d9921" containerName="dnsmasq-dns" containerID="cri-o://02fb7a6e347fb3a558e6cd6a6dbef20648452461f42bf80aff0be2b8bfc262be" gracePeriod=10 Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.800713 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7zfdn"] Oct 07 22:29:03 crc kubenswrapper[4871]: I1007 22:29:03.893235 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7sq86"] Oct 07 22:29:04 crc kubenswrapper[4871]: W1007 22:29:04.046506 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22863eb4_bd2f_42d6_a60c_5489bf5c1c81.slice/crio-6d640417b45a35e4c3c09b61f4e90302fc3050a1e989fa94cd64468be1ae8eac WatchSource:0}: Error finding container 6d640417b45a35e4c3c09b61f4e90302fc3050a1e989fa94cd64468be1ae8eac: Status 404 returned error can't find the container with id 6d640417b45a35e4c3c09b61f4e90302fc3050a1e989fa94cd64468be1ae8eac Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.306225 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerStarted","Data":"eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46"} Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.309518 4871 generic.go:334] "Generic (PLEG): container finished" podID="63e41063-fa11-4824-87df-49c82c6d9921" containerID="02fb7a6e347fb3a558e6cd6a6dbef20648452461f42bf80aff0be2b8bfc262be" exitCode=0 Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.309613 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" event={"ID":"63e41063-fa11-4824-87df-49c82c6d9921","Type":"ContainerDied","Data":"02fb7a6e347fb3a558e6cd6a6dbef20648452461f42bf80aff0be2b8bfc262be"} Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.310838 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7sq86" event={"ID":"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6","Type":"ContainerStarted","Data":"d74397ddaf67e27d6da2f67fa5ff091b2f2213bb519a7766c93af1c92ccb7aef"} Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.313728 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7zfdn" event={"ID":"22863eb4-bd2f-42d6-a60c-5489bf5c1c81","Type":"ContainerStarted","Data":"6d640417b45a35e4c3c09b61f4e90302fc3050a1e989fa94cd64468be1ae8eac"} Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.313843 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.313855 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.376291 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.414579 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.422161 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.433316 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.444877 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:29:04 crc kubenswrapper[4871]: E1007 22:29:04.445476 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerName="glance-httpd" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.445499 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerName="glance-httpd" Oct 07 22:29:04 crc kubenswrapper[4871]: E1007 22:29:04.445545 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f1c976-e97d-40b9-a97f-105920655bda" containerName="glance-log" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.445555 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f1c976-e97d-40b9-a97f-105920655bda" containerName="glance-log" Oct 07 22:29:04 crc kubenswrapper[4871]: E1007 22:29:04.445565 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerName="glance-log" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.445572 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerName="glance-log" Oct 07 22:29:04 crc kubenswrapper[4871]: E1007 22:29:04.445589 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f1c976-e97d-40b9-a97f-105920655bda" containerName="glance-httpd" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.445598 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f1c976-e97d-40b9-a97f-105920655bda" containerName="glance-httpd" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.445785 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f1c976-e97d-40b9-a97f-105920655bda" containerName="glance-log" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.445815 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f1c976-e97d-40b9-a97f-105920655bda" containerName="glance-httpd" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.445840 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerName="glance-log" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.445855 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9836b5be-f2ba-41fa-81aa-930b045e81f7" containerName="glance-httpd" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.447020 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.453559 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.453902 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.454276 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.454597 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8nrnc" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.454623 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.463059 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.464845 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.467559 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.470034 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.480116 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.541500 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.541554 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.541591 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-scripts\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.541667 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxms8\" (UniqueName: \"kubernetes.io/projected/537c4362-75e4-4c06-afff-55499ffed456-kube-api-access-vxms8\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.541748 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-logs\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.541881 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542013 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-config-data\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542085 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-scripts\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542112 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvcwv\" (UniqueName: \"kubernetes.io/projected/fb34122e-91f6-4318-8666-f85ea6fd8fbc-kube-api-access-dvcwv\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542143 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542191 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542444 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542569 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-logs\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542629 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542772 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-config-data\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.542863 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.573291 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.644804 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.644894 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.644928 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.644957 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-logs\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.644978 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645014 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-config-data\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645041 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645070 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645092 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645111 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-scripts\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645150 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxms8\" (UniqueName: \"kubernetes.io/projected/537c4362-75e4-4c06-afff-55499ffed456-kube-api-access-vxms8\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645178 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-logs\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645199 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645241 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-config-data\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645279 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-scripts\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.645300 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvcwv\" (UniqueName: \"kubernetes.io/projected/fb34122e-91f6-4318-8666-f85ea6fd8fbc-kube-api-access-dvcwv\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.646742 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-logs\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.646785 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-logs\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.646832 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.647082 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.647145 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.647756 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.656235 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.662076 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-scripts\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.670998 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-scripts\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.682958 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.684476 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-config-data\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.690203 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.678057 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.698288 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxms8\" (UniqueName: \"kubernetes.io/projected/537c4362-75e4-4c06-afff-55499ffed456-kube-api-access-vxms8\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.701172 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvcwv\" (UniqueName: \"kubernetes.io/projected/fb34122e-91f6-4318-8666-f85ea6fd8fbc-kube-api-access-dvcwv\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.706567 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-config-data\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.746851 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-sb\") pod \"63e41063-fa11-4824-87df-49c82c6d9921\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.746936 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m452\" (UniqueName: \"kubernetes.io/projected/63e41063-fa11-4824-87df-49c82c6d9921-kube-api-access-7m452\") pod \"63e41063-fa11-4824-87df-49c82c6d9921\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.747410 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-nb\") pod \"63e41063-fa11-4824-87df-49c82c6d9921\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.747593 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-svc\") pod \"63e41063-fa11-4824-87df-49c82c6d9921\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.747650 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-config\") pod \"63e41063-fa11-4824-87df-49c82c6d9921\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.747700 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-swift-storage-0\") pod \"63e41063-fa11-4824-87df-49c82c6d9921\" (UID: \"63e41063-fa11-4824-87df-49c82c6d9921\") " Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.756780 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.760414 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63e41063-fa11-4824-87df-49c82c6d9921-kube-api-access-7m452" (OuterVolumeSpecName: "kube-api-access-7m452") pod "63e41063-fa11-4824-87df-49c82c6d9921" (UID: "63e41063-fa11-4824-87df-49c82c6d9921"). InnerVolumeSpecName "kube-api-access-7m452". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.761211 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.807541 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "63e41063-fa11-4824-87df-49c82c6d9921" (UID: "63e41063-fa11-4824-87df-49c82c6d9921"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.810995 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "63e41063-fa11-4824-87df-49c82c6d9921" (UID: "63e41063-fa11-4824-87df-49c82c6d9921"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.823335 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-config" (OuterVolumeSpecName: "config") pod "63e41063-fa11-4824-87df-49c82c6d9921" (UID: "63e41063-fa11-4824-87df-49c82c6d9921"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.826239 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63e41063-fa11-4824-87df-49c82c6d9921" (UID: "63e41063-fa11-4824-87df-49c82c6d9921"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.848252 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63e41063-fa11-4824-87df-49c82c6d9921" (UID: "63e41063-fa11-4824-87df-49c82c6d9921"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.853058 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.857372 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.857412 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.857425 4871 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.857439 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.857449 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m452\" (UniqueName: \"kubernetes.io/projected/63e41063-fa11-4824-87df-49c82c6d9921-kube-api-access-7m452\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.857457 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e41063-fa11-4824-87df-49c82c6d9921-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.874345 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.877316 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ptrx5"] Oct 07 22:29:04 crc kubenswrapper[4871]: E1007 22:29:04.880882 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e41063-fa11-4824-87df-49c82c6d9921" containerName="dnsmasq-dns" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.880920 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e41063-fa11-4824-87df-49c82c6d9921" containerName="dnsmasq-dns" Oct 07 22:29:04 crc kubenswrapper[4871]: E1007 22:29:04.880975 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e41063-fa11-4824-87df-49c82c6d9921" containerName="init" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.880982 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e41063-fa11-4824-87df-49c82c6d9921" containerName="init" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.886180 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="63e41063-fa11-4824-87df-49c82c6d9921" containerName="dnsmasq-dns" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.887060 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.893832 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.894089 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2fh5m" Oct 07 22:29:04 crc kubenswrapper[4871]: I1007 22:29:04.894985 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ptrx5"] Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.010397 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9836b5be-f2ba-41fa-81aa-930b045e81f7" path="/var/lib/kubelet/pods/9836b5be-f2ba-41fa-81aa-930b045e81f7/volumes" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.011787 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0f1c976-e97d-40b9-a97f-105920655bda" path="/var/lib/kubelet/pods/d0f1c976-e97d-40b9-a97f-105920655bda/volumes" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.012750 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4lp7t"] Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.015956 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4lp7t"] Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.016081 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.022908 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.023328 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.023915 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5s9n4" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.060634 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4ws8\" (UniqueName: \"kubernetes.io/projected/88482765-93f6-4458-8511-1d932ebf2e16-kube-api-access-j4ws8\") pod \"barbican-db-sync-ptrx5\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.060712 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-db-sync-config-data\") pod \"barbican-db-sync-ptrx5\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.060804 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-combined-ca-bundle\") pod \"barbican-db-sync-ptrx5\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.164926 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4ws8\" (UniqueName: \"kubernetes.io/projected/88482765-93f6-4458-8511-1d932ebf2e16-kube-api-access-j4ws8\") pod \"barbican-db-sync-ptrx5\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.165526 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zjjh\" (UniqueName: \"kubernetes.io/projected/3dea95fc-3d17-44db-bece-3bf7deaf387a-kube-api-access-8zjjh\") pod \"neutron-db-sync-4lp7t\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.165580 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-db-sync-config-data\") pod \"barbican-db-sync-ptrx5\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.165695 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-combined-ca-bundle\") pod \"neutron-db-sync-4lp7t\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.165825 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-combined-ca-bundle\") pod \"barbican-db-sync-ptrx5\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.165921 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-config\") pod \"neutron-db-sync-4lp7t\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.174482 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-db-sync-config-data\") pod \"barbican-db-sync-ptrx5\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.184382 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-combined-ca-bundle\") pod \"barbican-db-sync-ptrx5\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.184407 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4ws8\" (UniqueName: \"kubernetes.io/projected/88482765-93f6-4458-8511-1d932ebf2e16-kube-api-access-j4ws8\") pod \"barbican-db-sync-ptrx5\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.228423 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.268474 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zjjh\" (UniqueName: \"kubernetes.io/projected/3dea95fc-3d17-44db-bece-3bf7deaf387a-kube-api-access-8zjjh\") pod \"neutron-db-sync-4lp7t\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.268549 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-combined-ca-bundle\") pod \"neutron-db-sync-4lp7t\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.268897 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-config\") pod \"neutron-db-sync-4lp7t\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.276705 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-config\") pod \"neutron-db-sync-4lp7t\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.287995 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-combined-ca-bundle\") pod \"neutron-db-sync-4lp7t\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.293449 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zjjh\" (UniqueName: \"kubernetes.io/projected/3dea95fc-3d17-44db-bece-3bf7deaf387a-kube-api-access-8zjjh\") pod \"neutron-db-sync-4lp7t\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.342400 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" event={"ID":"63e41063-fa11-4824-87df-49c82c6d9921","Type":"ContainerDied","Data":"cc76a46d042345982a49ccdcac3ea839c0b32425bddd8f66ae983ca2aecfecee"} Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.342476 4871 scope.go:117] "RemoveContainer" containerID="02fb7a6e347fb3a558e6cd6a6dbef20648452461f42bf80aff0be2b8bfc262be" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.342666 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f45f6cf7-f26k9" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.352580 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7zfdn" event={"ID":"22863eb4-bd2f-42d6-a60c-5489bf5c1c81","Type":"ContainerStarted","Data":"9ae44c9c6b44d826a4bf86d7e4c7289e841ebde0a866bf0d9aa672e7b5e7c0bd"} Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.380232 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-f26k9"] Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.395413 4871 scope.go:117] "RemoveContainer" containerID="4ef6f7552387c8d751270f4aee8227aba0784289050826a18691cd405931e855" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.409009 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-f26k9"] Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.417605 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-7zfdn" podStartSLOduration=4.417580611 podStartE2EDuration="4.417580611s" podCreationTimestamp="2025-10-07 22:29:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:05.390179294 +0000 UTC m=+1219.192877367" watchObservedRunningTime="2025-10-07 22:29:05.417580611 +0000 UTC m=+1219.220278684" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.555600 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.573761 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:29:05 crc kubenswrapper[4871]: I1007 22:29:05.807875 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ptrx5"] Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.058814 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4lp7t"] Oct 07 22:29:06 crc kubenswrapper[4871]: W1007 22:29:06.068059 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dea95fc_3d17_44db_bece_3bf7deaf387a.slice/crio-b1935d4bc5699c75a2b8f21ae7f21a3243b822b98acf6c4636fee93345c9c97a WatchSource:0}: Error finding container b1935d4bc5699c75a2b8f21ae7f21a3243b822b98acf6c4636fee93345c9c97a: Status 404 returned error can't find the container with id b1935d4bc5699c75a2b8f21ae7f21a3243b822b98acf6c4636fee93345c9c97a Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.091814 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.379852 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb34122e-91f6-4318-8666-f85ea6fd8fbc","Type":"ContainerStarted","Data":"d433d98e82a8966ce766449f7bdea2d5cc7e768065f9bb55a27556194b51bd6b"} Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.388162 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4lp7t" event={"ID":"3dea95fc-3d17-44db-bece-3bf7deaf387a","Type":"ContainerStarted","Data":"673a2fb678380df6b3c6dd2cb9e872c1ff636cea1ee56f22ae5bf5386ab04451"} Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.388198 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4lp7t" event={"ID":"3dea95fc-3d17-44db-bece-3bf7deaf387a","Type":"ContainerStarted","Data":"b1935d4bc5699c75a2b8f21ae7f21a3243b822b98acf6c4636fee93345c9c97a"} Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.389834 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ptrx5" event={"ID":"88482765-93f6-4458-8511-1d932ebf2e16","Type":"ContainerStarted","Data":"aefeb9e7163fc3232c96f8e14e45fc048e6fc4f47fb1d7175e0007451ed9e5b1"} Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.400112 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zt6f9" event={"ID":"f0b7dc36-3f82-4650-b926-03de75dda7e4","Type":"ContainerStarted","Data":"fa8c238c4e4dd8b2b2f155c5c018100d3ed579f7df72b59d5814579618e15651"} Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.410008 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"537c4362-75e4-4c06-afff-55499ffed456","Type":"ContainerStarted","Data":"8dd9a28f4c6c9641ac6b27e2e4d221635516407ec09734b7f97b3fdb49856b47"} Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.410072 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"537c4362-75e4-4c06-afff-55499ffed456","Type":"ContainerStarted","Data":"601b59ada2c4a8f358447ac89f74f115ebceff7fa3a0899001ad2ccb3838426f"} Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.410748 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4lp7t" podStartSLOduration=2.410712952 podStartE2EDuration="2.410712952s" podCreationTimestamp="2025-10-07 22:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:06.407184238 +0000 UTC m=+1220.209882321" watchObservedRunningTime="2025-10-07 22:29:06.410712952 +0000 UTC m=+1220.213411015" Oct 07 22:29:06 crc kubenswrapper[4871]: I1007 22:29:06.458931 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-zt6f9" podStartSLOduration=2.700146674 podStartE2EDuration="13.45890816s" podCreationTimestamp="2025-10-07 22:28:53 +0000 UTC" firstStartedPulling="2025-10-07 22:28:54.163664707 +0000 UTC m=+1207.966362780" lastFinishedPulling="2025-10-07 22:29:04.922426193 +0000 UTC m=+1218.725124266" observedRunningTime="2025-10-07 22:29:06.440633895 +0000 UTC m=+1220.243331958" watchObservedRunningTime="2025-10-07 22:29:06.45890816 +0000 UTC m=+1220.261606233" Oct 07 22:29:07 crc kubenswrapper[4871]: I1007 22:29:07.000704 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63e41063-fa11-4824-87df-49c82c6d9921" path="/var/lib/kubelet/pods/63e41063-fa11-4824-87df-49c82c6d9921/volumes" Oct 07 22:29:07 crc kubenswrapper[4871]: I1007 22:29:07.421407 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerStarted","Data":"0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed"} Oct 07 22:29:07 crc kubenswrapper[4871]: I1007 22:29:07.425832 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb34122e-91f6-4318-8666-f85ea6fd8fbc","Type":"ContainerStarted","Data":"029c5ef52594f3600fed24649c374ab278a7e185161657daa6156f34dffcf1a9"} Oct 07 22:29:08 crc kubenswrapper[4871]: I1007 22:29:08.442821 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb34122e-91f6-4318-8666-f85ea6fd8fbc","Type":"ContainerStarted","Data":"17ad90817b0347a5b221ac9e0855d33dc50476e984f82363e5299dd2e53ebf52"} Oct 07 22:29:08 crc kubenswrapper[4871]: I1007 22:29:08.446341 4871 generic.go:334] "Generic (PLEG): container finished" podID="f0b7dc36-3f82-4650-b926-03de75dda7e4" containerID="fa8c238c4e4dd8b2b2f155c5c018100d3ed579f7df72b59d5814579618e15651" exitCode=0 Oct 07 22:29:08 crc kubenswrapper[4871]: I1007 22:29:08.446432 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zt6f9" event={"ID":"f0b7dc36-3f82-4650-b926-03de75dda7e4","Type":"ContainerDied","Data":"fa8c238c4e4dd8b2b2f155c5c018100d3ed579f7df72b59d5814579618e15651"} Oct 07 22:29:08 crc kubenswrapper[4871]: I1007 22:29:08.451130 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"537c4362-75e4-4c06-afff-55499ffed456","Type":"ContainerStarted","Data":"8173a9d1525d122748f48123b0201f7fa734f3da4bc057a26edb4c400c38a339"} Oct 07 22:29:08 crc kubenswrapper[4871]: I1007 22:29:08.471347 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.471317094 podStartE2EDuration="4.471317094s" podCreationTimestamp="2025-10-07 22:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:08.470601915 +0000 UTC m=+1222.273299988" watchObservedRunningTime="2025-10-07 22:29:08.471317094 +0000 UTC m=+1222.274015167" Oct 07 22:29:08 crc kubenswrapper[4871]: I1007 22:29:08.508365 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.508339576 podStartE2EDuration="4.508339576s" podCreationTimestamp="2025-10-07 22:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:08.492509356 +0000 UTC m=+1222.295207429" watchObservedRunningTime="2025-10-07 22:29:08.508339576 +0000 UTC m=+1222.311037649" Oct 07 22:29:09 crc kubenswrapper[4871]: I1007 22:29:09.467331 4871 generic.go:334] "Generic (PLEG): container finished" podID="22863eb4-bd2f-42d6-a60c-5489bf5c1c81" containerID="9ae44c9c6b44d826a4bf86d7e4c7289e841ebde0a866bf0d9aa672e7b5e7c0bd" exitCode=0 Oct 07 22:29:09 crc kubenswrapper[4871]: I1007 22:29:09.467445 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7zfdn" event={"ID":"22863eb4-bd2f-42d6-a60c-5489bf5c1c81","Type":"ContainerDied","Data":"9ae44c9c6b44d826a4bf86d7e4c7289e841ebde0a866bf0d9aa672e7b5e7c0bd"} Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.555376 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zt6f9" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.574619 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zt6f9" event={"ID":"f0b7dc36-3f82-4650-b926-03de75dda7e4","Type":"ContainerDied","Data":"bf6cb3e378505b49aab3cf0349e14ec411e6b52e50a4f92bf77b3f8fb42bdae5"} Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.574667 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf6cb3e378505b49aab3cf0349e14ec411e6b52e50a4f92bf77b3f8fb42bdae5" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.641295 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-scripts\") pod \"f0b7dc36-3f82-4650-b926-03de75dda7e4\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.641532 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-combined-ca-bundle\") pod \"f0b7dc36-3f82-4650-b926-03de75dda7e4\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.641607 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b7dc36-3f82-4650-b926-03de75dda7e4-logs\") pod \"f0b7dc36-3f82-4650-b926-03de75dda7e4\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.641656 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-config-data\") pod \"f0b7dc36-3f82-4650-b926-03de75dda7e4\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.641712 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db9tz\" (UniqueName: \"kubernetes.io/projected/f0b7dc36-3f82-4650-b926-03de75dda7e4-kube-api-access-db9tz\") pod \"f0b7dc36-3f82-4650-b926-03de75dda7e4\" (UID: \"f0b7dc36-3f82-4650-b926-03de75dda7e4\") " Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.642526 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0b7dc36-3f82-4650-b926-03de75dda7e4-logs" (OuterVolumeSpecName: "logs") pod "f0b7dc36-3f82-4650-b926-03de75dda7e4" (UID: "f0b7dc36-3f82-4650-b926-03de75dda7e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.656934 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-scripts" (OuterVolumeSpecName: "scripts") pod "f0b7dc36-3f82-4650-b926-03de75dda7e4" (UID: "f0b7dc36-3f82-4650-b926-03de75dda7e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.675362 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0b7dc36-3f82-4650-b926-03de75dda7e4" (UID: "f0b7dc36-3f82-4650-b926-03de75dda7e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.679604 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0b7dc36-3f82-4650-b926-03de75dda7e4-kube-api-access-db9tz" (OuterVolumeSpecName: "kube-api-access-db9tz") pod "f0b7dc36-3f82-4650-b926-03de75dda7e4" (UID: "f0b7dc36-3f82-4650-b926-03de75dda7e4"). InnerVolumeSpecName "kube-api-access-db9tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.695904 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-config-data" (OuterVolumeSpecName: "config-data") pod "f0b7dc36-3f82-4650-b926-03de75dda7e4" (UID: "f0b7dc36-3f82-4650-b926-03de75dda7e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.743541 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db9tz\" (UniqueName: \"kubernetes.io/projected/f0b7dc36-3f82-4650-b926-03de75dda7e4-kube-api-access-db9tz\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.743588 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.743602 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.743611 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b7dc36-3f82-4650-b926-03de75dda7e4-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:12 crc kubenswrapper[4871]: I1007 22:29:12.743619 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b7dc36-3f82-4650-b926-03de75dda7e4-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.585521 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zt6f9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.680459 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7567f46b5d-kb4h9"] Oct 07 22:29:13 crc kubenswrapper[4871]: E1007 22:29:13.680833 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b7dc36-3f82-4650-b926-03de75dda7e4" containerName="placement-db-sync" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.680852 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b7dc36-3f82-4650-b926-03de75dda7e4" containerName="placement-db-sync" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.681043 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0b7dc36-3f82-4650-b926-03de75dda7e4" containerName="placement-db-sync" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.681887 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.692435 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.692675 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bqwbw" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.693203 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.693250 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.695571 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.707198 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7567f46b5d-kb4h9"] Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.774396 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-internal-tls-certs\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.774907 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-public-tls-certs\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.774974 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-config-data\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.775111 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/878958ac-64b8-434b-88ee-f23d3c486765-logs\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.775277 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-combined-ca-bundle\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.775448 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79bzm\" (UniqueName: \"kubernetes.io/projected/878958ac-64b8-434b-88ee-f23d3c486765-kube-api-access-79bzm\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.775510 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-scripts\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.878499 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-internal-tls-certs\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.878568 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-public-tls-certs\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.878608 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-config-data\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.878632 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/878958ac-64b8-434b-88ee-f23d3c486765-logs\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.878673 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-combined-ca-bundle\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.878718 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79bzm\" (UniqueName: \"kubernetes.io/projected/878958ac-64b8-434b-88ee-f23d3c486765-kube-api-access-79bzm\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.878741 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-scripts\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.882392 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/878958ac-64b8-434b-88ee-f23d3c486765-logs\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.891902 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-scripts\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.892527 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-public-tls-certs\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.892703 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-combined-ca-bundle\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.893246 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-config-data\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.902835 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79bzm\" (UniqueName: \"kubernetes.io/projected/878958ac-64b8-434b-88ee-f23d3c486765-kube-api-access-79bzm\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:13 crc kubenswrapper[4871]: I1007 22:29:13.910651 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-internal-tls-certs\") pod \"placement-7567f46b5d-kb4h9\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:14 crc kubenswrapper[4871]: I1007 22:29:14.014825 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:14 crc kubenswrapper[4871]: I1007 22:29:14.854360 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:14 crc kubenswrapper[4871]: I1007 22:29:14.854431 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:14 crc kubenswrapper[4871]: I1007 22:29:14.875993 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 22:29:14 crc kubenswrapper[4871]: I1007 22:29:14.876079 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 22:29:14 crc kubenswrapper[4871]: I1007 22:29:14.912763 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:15 crc kubenswrapper[4871]: I1007 22:29:15.046688 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 22:29:15 crc kubenswrapper[4871]: I1007 22:29:15.068931 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 22:29:15 crc kubenswrapper[4871]: I1007 22:29:15.116223 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:15 crc kubenswrapper[4871]: I1007 22:29:15.610131 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:15 crc kubenswrapper[4871]: I1007 22:29:15.610204 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 22:29:15 crc kubenswrapper[4871]: I1007 22:29:15.610231 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 22:29:15 crc kubenswrapper[4871]: I1007 22:29:15.610249 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.027629 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.165259 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-combined-ca-bundle\") pod \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.165375 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-credential-keys\") pod \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.165426 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-scripts\") pod \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.165481 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-fernet-keys\") pod \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.165506 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-config-data\") pod \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.165553 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwzcp\" (UniqueName: \"kubernetes.io/projected/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-kube-api-access-kwzcp\") pod \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\" (UID: \"22863eb4-bd2f-42d6-a60c-5489bf5c1c81\") " Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.178998 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "22863eb4-bd2f-42d6-a60c-5489bf5c1c81" (UID: "22863eb4-bd2f-42d6-a60c-5489bf5c1c81"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.179053 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-scripts" (OuterVolumeSpecName: "scripts") pod "22863eb4-bd2f-42d6-a60c-5489bf5c1c81" (UID: "22863eb4-bd2f-42d6-a60c-5489bf5c1c81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.179103 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-kube-api-access-kwzcp" (OuterVolumeSpecName: "kube-api-access-kwzcp") pod "22863eb4-bd2f-42d6-a60c-5489bf5c1c81" (UID: "22863eb4-bd2f-42d6-a60c-5489bf5c1c81"). InnerVolumeSpecName "kube-api-access-kwzcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.181867 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "22863eb4-bd2f-42d6-a60c-5489bf5c1c81" (UID: "22863eb4-bd2f-42d6-a60c-5489bf5c1c81"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.204847 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-config-data" (OuterVolumeSpecName: "config-data") pod "22863eb4-bd2f-42d6-a60c-5489bf5c1c81" (UID: "22863eb4-bd2f-42d6-a60c-5489bf5c1c81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.228030 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22863eb4-bd2f-42d6-a60c-5489bf5c1c81" (UID: "22863eb4-bd2f-42d6-a60c-5489bf5c1c81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.268239 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.268315 4871 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.268330 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.268343 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwzcp\" (UniqueName: \"kubernetes.io/projected/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-kube-api-access-kwzcp\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.268354 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.268365 4871 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/22863eb4-bd2f-42d6-a60c-5489bf5c1c81-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.637032 4871 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.637053 4871 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.637322 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7zfdn" event={"ID":"22863eb4-bd2f-42d6-a60c-5489bf5c1c81","Type":"ContainerDied","Data":"6d640417b45a35e4c3c09b61f4e90302fc3050a1e989fa94cd64468be1ae8eac"} Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.637383 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7zfdn" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.637399 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d640417b45a35e4c3c09b61f4e90302fc3050a1e989fa94cd64468be1ae8eac" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.690637 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.780319 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.780765 4871 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.787345 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 22:29:17 crc kubenswrapper[4871]: I1007 22:29:17.822003 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.144180 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cd55cdc75-kj6jh"] Oct 07 22:29:18 crc kubenswrapper[4871]: E1007 22:29:18.144904 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22863eb4-bd2f-42d6-a60c-5489bf5c1c81" containerName="keystone-bootstrap" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.144922 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="22863eb4-bd2f-42d6-a60c-5489bf5c1c81" containerName="keystone-bootstrap" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.145157 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="22863eb4-bd2f-42d6-a60c-5489bf5c1c81" containerName="keystone-bootstrap" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.145893 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.162573 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.162966 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.165422 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.165645 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.166697 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4kmth" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.168331 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.185370 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cd55cdc75-kj6jh"] Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.305469 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-scripts\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.305579 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-fernet-keys\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.305613 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-combined-ca-bundle\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.305635 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t7gq\" (UniqueName: \"kubernetes.io/projected/6af0bcb3-3803-4174-af06-c6841610c62a-kube-api-access-4t7gq\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.305665 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-credential-keys\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.305731 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-public-tls-certs\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.305757 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-config-data\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.305817 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-internal-tls-certs\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.407665 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-credential-keys\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.407775 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-public-tls-certs\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.407833 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-config-data\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.407879 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-internal-tls-certs\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.407922 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-scripts\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.407996 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-fernet-keys\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.408029 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-combined-ca-bundle\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.408056 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t7gq\" (UniqueName: \"kubernetes.io/projected/6af0bcb3-3803-4174-af06-c6841610c62a-kube-api-access-4t7gq\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.417075 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-scripts\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.417631 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-internal-tls-certs\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.419647 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-fernet-keys\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.420174 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-public-tls-certs\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.421556 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-credential-keys\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.425496 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-config-data\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.427361 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-combined-ca-bundle\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.434537 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t7gq\" (UniqueName: \"kubernetes.io/projected/6af0bcb3-3803-4174-af06-c6841610c62a-kube-api-access-4t7gq\") pod \"keystone-cd55cdc75-kj6jh\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:18 crc kubenswrapper[4871]: I1007 22:29:18.476635 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:26 crc kubenswrapper[4871]: E1007 22:29:26.146149 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dea95fc_3d17_44db_bece_3bf7deaf387a.slice/crio-conmon-673a2fb678380df6b3c6dd2cb9e872c1ff636cea1ee56f22ae5bf5386ab04451.scope\": RecentStats: unable to find data in memory cache]" Oct 07 22:29:26 crc kubenswrapper[4871]: I1007 22:29:26.749053 4871 generic.go:334] "Generic (PLEG): container finished" podID="3dea95fc-3d17-44db-bece-3bf7deaf387a" containerID="673a2fb678380df6b3c6dd2cb9e872c1ff636cea1ee56f22ae5bf5386ab04451" exitCode=0 Oct 07 22:29:26 crc kubenswrapper[4871]: I1007 22:29:26.749132 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4lp7t" event={"ID":"3dea95fc-3d17-44db-bece-3bf7deaf387a","Type":"ContainerDied","Data":"673a2fb678380df6b3c6dd2cb9e872c1ff636cea1ee56f22ae5bf5386ab04451"} Oct 07 22:29:27 crc kubenswrapper[4871]: E1007 22:29:27.585716 4871 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f" Oct 07 22:29:27 crc kubenswrapper[4871]: E1007 22:29:27.586487 4871 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hps9v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7sq86_openstack(a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 22:29:27 crc kubenswrapper[4871]: E1007 22:29:27.587884 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7sq86" podUID="a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" Oct 07 22:29:27 crc kubenswrapper[4871]: E1007 22:29:27.775686 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f\\\"\"" pod="openstack/cinder-db-sync-7sq86" podUID="a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.109636 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7567f46b5d-kb4h9"] Oct 07 22:29:28 crc kubenswrapper[4871]: W1007 22:29:28.112704 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod878958ac_64b8_434b_88ee_f23d3c486765.slice/crio-dd20e5029ebb0b145df9d9a14d23475bba65506ab1c015e2e51ef4f3933650bd WatchSource:0}: Error finding container dd20e5029ebb0b145df9d9a14d23475bba65506ab1c015e2e51ef4f3933650bd: Status 404 returned error can't find the container with id dd20e5029ebb0b145df9d9a14d23475bba65506ab1c015e2e51ef4f3933650bd Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.159796 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cd55cdc75-kj6jh"] Oct 07 22:29:28 crc kubenswrapper[4871]: W1007 22:29:28.166648 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6af0bcb3_3803_4174_af06_c6841610c62a.slice/crio-72047d6988570f2778705a14820615c6601e749d79b7cf666481ed8423c52cb0 WatchSource:0}: Error finding container 72047d6988570f2778705a14820615c6601e749d79b7cf666481ed8423c52cb0: Status 404 returned error can't find the container with id 72047d6988570f2778705a14820615c6601e749d79b7cf666481ed8423c52cb0 Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.202191 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.279148 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-combined-ca-bundle\") pod \"3dea95fc-3d17-44db-bece-3bf7deaf387a\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.279927 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zjjh\" (UniqueName: \"kubernetes.io/projected/3dea95fc-3d17-44db-bece-3bf7deaf387a-kube-api-access-8zjjh\") pod \"3dea95fc-3d17-44db-bece-3bf7deaf387a\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.280318 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-config\") pod \"3dea95fc-3d17-44db-bece-3bf7deaf387a\" (UID: \"3dea95fc-3d17-44db-bece-3bf7deaf387a\") " Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.284420 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dea95fc-3d17-44db-bece-3bf7deaf387a-kube-api-access-8zjjh" (OuterVolumeSpecName: "kube-api-access-8zjjh") pod "3dea95fc-3d17-44db-bece-3bf7deaf387a" (UID: "3dea95fc-3d17-44db-bece-3bf7deaf387a"). InnerVolumeSpecName "kube-api-access-8zjjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.316080 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dea95fc-3d17-44db-bece-3bf7deaf387a" (UID: "3dea95fc-3d17-44db-bece-3bf7deaf387a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.318269 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-config" (OuterVolumeSpecName: "config") pod "3dea95fc-3d17-44db-bece-3bf7deaf387a" (UID: "3dea95fc-3d17-44db-bece-3bf7deaf387a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.383106 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.383144 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zjjh\" (UniqueName: \"kubernetes.io/projected/3dea95fc-3d17-44db-bece-3bf7deaf387a-kube-api-access-8zjjh\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.383159 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3dea95fc-3d17-44db-bece-3bf7deaf387a-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.780712 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerStarted","Data":"b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38"} Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.782278 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4lp7t" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.782270 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4lp7t" event={"ID":"3dea95fc-3d17-44db-bece-3bf7deaf387a","Type":"ContainerDied","Data":"b1935d4bc5699c75a2b8f21ae7f21a3243b822b98acf6c4636fee93345c9c97a"} Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.782405 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1935d4bc5699c75a2b8f21ae7f21a3243b822b98acf6c4636fee93345c9c97a" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.783536 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ptrx5" event={"ID":"88482765-93f6-4458-8511-1d932ebf2e16","Type":"ContainerStarted","Data":"e9b84dbda4c0fa3402e53853a13275e7a4791285459387786a2900afc2dbb3a7"} Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.786300 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cd55cdc75-kj6jh" event={"ID":"6af0bcb3-3803-4174-af06-c6841610c62a","Type":"ContainerStarted","Data":"1987ababc57bf8086a0c131045faa2ea781c10edf009a458861520b942f65314"} Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.786331 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cd55cdc75-kj6jh" event={"ID":"6af0bcb3-3803-4174-af06-c6841610c62a","Type":"ContainerStarted","Data":"72047d6988570f2778705a14820615c6601e749d79b7cf666481ed8423c52cb0"} Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.786778 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.788210 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7567f46b5d-kb4h9" event={"ID":"878958ac-64b8-434b-88ee-f23d3c486765","Type":"ContainerStarted","Data":"858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c"} Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.788262 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7567f46b5d-kb4h9" event={"ID":"878958ac-64b8-434b-88ee-f23d3c486765","Type":"ContainerStarted","Data":"ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc"} Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.788277 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7567f46b5d-kb4h9" event={"ID":"878958ac-64b8-434b-88ee-f23d3c486765","Type":"ContainerStarted","Data":"dd20e5029ebb0b145df9d9a14d23475bba65506ab1c015e2e51ef4f3933650bd"} Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.788942 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.788973 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.864334 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ptrx5" podStartSLOduration=3.159188486 podStartE2EDuration="24.864310893s" podCreationTimestamp="2025-10-07 22:29:04 +0000 UTC" firstStartedPulling="2025-10-07 22:29:05.827849138 +0000 UTC m=+1219.630547211" lastFinishedPulling="2025-10-07 22:29:27.532971535 +0000 UTC m=+1241.335669618" observedRunningTime="2025-10-07 22:29:28.818363065 +0000 UTC m=+1242.621061148" watchObservedRunningTime="2025-10-07 22:29:28.864310893 +0000 UTC m=+1242.667008966" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.877946 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7567f46b5d-kb4h9" podStartSLOduration=15.877918634 podStartE2EDuration="15.877918634s" podCreationTimestamp="2025-10-07 22:29:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:28.847903478 +0000 UTC m=+1242.650601551" watchObservedRunningTime="2025-10-07 22:29:28.877918634 +0000 UTC m=+1242.680616707" Oct 07 22:29:28 crc kubenswrapper[4871]: I1007 22:29:28.884283 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cd55cdc75-kj6jh" podStartSLOduration=10.884261482 podStartE2EDuration="10.884261482s" podCreationTimestamp="2025-10-07 22:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:28.875609182 +0000 UTC m=+1242.678307255" watchObservedRunningTime="2025-10-07 22:29:28.884261482 +0000 UTC m=+1242.686959555" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.014910 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6ffd8547bc-kfdwq"] Oct 07 22:29:29 crc kubenswrapper[4871]: E1007 22:29:29.015439 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dea95fc-3d17-44db-bece-3bf7deaf387a" containerName="neutron-db-sync" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.015456 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dea95fc-3d17-44db-bece-3bf7deaf387a" containerName="neutron-db-sync" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.015646 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dea95fc-3d17-44db-bece-3bf7deaf387a" containerName="neutron-db-sync" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.016691 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.038132 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffd8547bc-kfdwq"] Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.113134 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-swift-storage-0\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.113267 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-svc\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.113350 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxdr5\" (UniqueName: \"kubernetes.io/projected/0cdd7757-e120-4571-87e8-85538c0c50b0-kube-api-access-vxdr5\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.113390 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.113487 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-config\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.113510 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.139627 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-74676f67cb-m997h"] Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.141149 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.146067 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.146089 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.146277 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5s9n4" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.146283 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.188398 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-74676f67cb-m997h"] Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.214709 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-config\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.214769 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.214834 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glfqm\" (UniqueName: \"kubernetes.io/projected/c4635daf-677e-4d01-adff-95f9aa80fa7e-kube-api-access-glfqm\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.214856 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-httpd-config\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.214895 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-swift-storage-0\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.214919 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-config\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.214938 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-svc\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.214978 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-combined-ca-bundle\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.215004 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxdr5\" (UniqueName: \"kubernetes.io/projected/0cdd7757-e120-4571-87e8-85538c0c50b0-kube-api-access-vxdr5\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.215021 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.215053 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-ovndb-tls-certs\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.216014 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-config\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.216586 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.218005 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-swift-storage-0\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.218133 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-svc\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.218738 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.251848 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxdr5\" (UniqueName: \"kubernetes.io/projected/0cdd7757-e120-4571-87e8-85538c0c50b0-kube-api-access-vxdr5\") pod \"dnsmasq-dns-6ffd8547bc-kfdwq\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.317146 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glfqm\" (UniqueName: \"kubernetes.io/projected/c4635daf-677e-4d01-adff-95f9aa80fa7e-kube-api-access-glfqm\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.317197 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-httpd-config\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.318146 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-config\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.318220 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-combined-ca-bundle\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.318280 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-ovndb-tls-certs\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.324132 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-httpd-config\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.324675 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-config\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.324700 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-ovndb-tls-certs\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.329155 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-combined-ca-bundle\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.338361 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glfqm\" (UniqueName: \"kubernetes.io/projected/c4635daf-677e-4d01-adff-95f9aa80fa7e-kube-api-access-glfqm\") pod \"neutron-74676f67cb-m997h\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.369699 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.490074 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:29 crc kubenswrapper[4871]: I1007 22:29:29.893136 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffd8547bc-kfdwq"] Oct 07 22:29:29 crc kubenswrapper[4871]: W1007 22:29:29.898562 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cdd7757_e120_4571_87e8_85538c0c50b0.slice/crio-7984a611b854beaa53e0e794d0e0265c16c4a418efe40723276b6a387e6402d5 WatchSource:0}: Error finding container 7984a611b854beaa53e0e794d0e0265c16c4a418efe40723276b6a387e6402d5: Status 404 returned error can't find the container with id 7984a611b854beaa53e0e794d0e0265c16c4a418efe40723276b6a387e6402d5 Oct 07 22:29:30 crc kubenswrapper[4871]: I1007 22:29:30.426028 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-74676f67cb-m997h"] Oct 07 22:29:30 crc kubenswrapper[4871]: I1007 22:29:30.810919 4871 generic.go:334] "Generic (PLEG): container finished" podID="0cdd7757-e120-4571-87e8-85538c0c50b0" containerID="c034333d104008806b90644a8cd76a24d9025f0b6ef3580a98e3110ee7931157" exitCode=0 Oct 07 22:29:30 crc kubenswrapper[4871]: I1007 22:29:30.811156 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" event={"ID":"0cdd7757-e120-4571-87e8-85538c0c50b0","Type":"ContainerDied","Data":"c034333d104008806b90644a8cd76a24d9025f0b6ef3580a98e3110ee7931157"} Oct 07 22:29:30 crc kubenswrapper[4871]: I1007 22:29:30.811548 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" event={"ID":"0cdd7757-e120-4571-87e8-85538c0c50b0","Type":"ContainerStarted","Data":"7984a611b854beaa53e0e794d0e0265c16c4a418efe40723276b6a387e6402d5"} Oct 07 22:29:30 crc kubenswrapper[4871]: I1007 22:29:30.813827 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74676f67cb-m997h" event={"ID":"c4635daf-677e-4d01-adff-95f9aa80fa7e","Type":"ContainerStarted","Data":"8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105"} Oct 07 22:29:30 crc kubenswrapper[4871]: I1007 22:29:30.813861 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74676f67cb-m997h" event={"ID":"c4635daf-677e-4d01-adff-95f9aa80fa7e","Type":"ContainerStarted","Data":"f5f2feb3d9dc5a1c2284ad75b1ff747d1312a2abe5db3864e426981506ad64d9"} Oct 07 22:29:30 crc kubenswrapper[4871]: I1007 22:29:30.815541 4871 generic.go:334] "Generic (PLEG): container finished" podID="88482765-93f6-4458-8511-1d932ebf2e16" containerID="e9b84dbda4c0fa3402e53853a13275e7a4791285459387786a2900afc2dbb3a7" exitCode=0 Oct 07 22:29:30 crc kubenswrapper[4871]: I1007 22:29:30.816654 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ptrx5" event={"ID":"88482765-93f6-4458-8511-1d932ebf2e16","Type":"ContainerDied","Data":"e9b84dbda4c0fa3402e53853a13275e7a4791285459387786a2900afc2dbb3a7"} Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.551550 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7cc4f647bc-9lrsh"] Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.560175 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.563368 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.564172 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cc4f647bc-9lrsh"] Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.564208 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.695990 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-config\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.696068 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pmf8\" (UniqueName: \"kubernetes.io/projected/7116e0fa-30d9-4a8d-9a11-0d57908f268f-kube-api-access-5pmf8\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.696622 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-httpd-config\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.696687 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-ovndb-tls-certs\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.696903 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-internal-tls-certs\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.696955 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-combined-ca-bundle\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.696993 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-public-tls-certs\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.798974 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-httpd-config\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.799035 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-ovndb-tls-certs\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.799081 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-internal-tls-certs\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.799126 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-combined-ca-bundle\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.799159 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-public-tls-certs\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.799221 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-config\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.799240 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pmf8\" (UniqueName: \"kubernetes.io/projected/7116e0fa-30d9-4a8d-9a11-0d57908f268f-kube-api-access-5pmf8\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.807930 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-internal-tls-certs\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.808965 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-ovndb-tls-certs\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.809101 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-config\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.809995 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-httpd-config\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.810012 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-public-tls-certs\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.810990 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-combined-ca-bundle\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.817573 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pmf8\" (UniqueName: \"kubernetes.io/projected/7116e0fa-30d9-4a8d-9a11-0d57908f268f-kube-api-access-5pmf8\") pod \"neutron-7cc4f647bc-9lrsh\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.834260 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" event={"ID":"0cdd7757-e120-4571-87e8-85538c0c50b0","Type":"ContainerStarted","Data":"3e1eaadc61aed052ef0e52fa6bd17cd42ed8c865d84e7d580bf6f63987ba92fe"} Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.836250 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.843793 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74676f67cb-m997h" event={"ID":"c4635daf-677e-4d01-adff-95f9aa80fa7e","Type":"ContainerStarted","Data":"3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f"} Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.843951 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.864400 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" podStartSLOduration=3.8643669430000003 podStartE2EDuration="3.864366943s" podCreationTimestamp="2025-10-07 22:29:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:31.857060729 +0000 UTC m=+1245.659758812" watchObservedRunningTime="2025-10-07 22:29:31.864366943 +0000 UTC m=+1245.667065016" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.890472 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-74676f67cb-m997h" podStartSLOduration=2.890421114 podStartE2EDuration="2.890421114s" podCreationTimestamp="2025-10-07 22:29:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:31.880232714 +0000 UTC m=+1245.682930807" watchObservedRunningTime="2025-10-07 22:29:31.890421114 +0000 UTC m=+1245.693119187" Oct 07 22:29:31 crc kubenswrapper[4871]: I1007 22:29:31.905953 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.213804 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.335748 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-db-sync-config-data\") pod \"88482765-93f6-4458-8511-1d932ebf2e16\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.335954 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4ws8\" (UniqueName: \"kubernetes.io/projected/88482765-93f6-4458-8511-1d932ebf2e16-kube-api-access-j4ws8\") pod \"88482765-93f6-4458-8511-1d932ebf2e16\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.336153 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-combined-ca-bundle\") pod \"88482765-93f6-4458-8511-1d932ebf2e16\" (UID: \"88482765-93f6-4458-8511-1d932ebf2e16\") " Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.341843 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88482765-93f6-4458-8511-1d932ebf2e16-kube-api-access-j4ws8" (OuterVolumeSpecName: "kube-api-access-j4ws8") pod "88482765-93f6-4458-8511-1d932ebf2e16" (UID: "88482765-93f6-4458-8511-1d932ebf2e16"). InnerVolumeSpecName "kube-api-access-j4ws8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.342125 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "88482765-93f6-4458-8511-1d932ebf2e16" (UID: "88482765-93f6-4458-8511-1d932ebf2e16"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.366798 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88482765-93f6-4458-8511-1d932ebf2e16" (UID: "88482765-93f6-4458-8511-1d932ebf2e16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.438643 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.438686 4871 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88482765-93f6-4458-8511-1d932ebf2e16-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.438702 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4ws8\" (UniqueName: \"kubernetes.io/projected/88482765-93f6-4458-8511-1d932ebf2e16-kube-api-access-j4ws8\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.613799 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cc4f647bc-9lrsh"] Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.858342 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cc4f647bc-9lrsh" event={"ID":"7116e0fa-30d9-4a8d-9a11-0d57908f268f","Type":"ContainerStarted","Data":"4d4fd8eb0e3f743f4a495e4fd83d8474249ee2494cd13c2d79f583b0d0386669"} Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.861205 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ptrx5" Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.861284 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ptrx5" event={"ID":"88482765-93f6-4458-8511-1d932ebf2e16","Type":"ContainerDied","Data":"aefeb9e7163fc3232c96f8e14e45fc048e6fc4f47fb1d7175e0007451ed9e5b1"} Oct 07 22:29:32 crc kubenswrapper[4871]: I1007 22:29:32.861328 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aefeb9e7163fc3232c96f8e14e45fc048e6fc4f47fb1d7175e0007451ed9e5b1" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.230456 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5544d45d85-dqp9x"] Oct 07 22:29:33 crc kubenswrapper[4871]: E1007 22:29:33.231108 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88482765-93f6-4458-8511-1d932ebf2e16" containerName="barbican-db-sync" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.231128 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="88482765-93f6-4458-8511-1d932ebf2e16" containerName="barbican-db-sync" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.231304 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="88482765-93f6-4458-8511-1d932ebf2e16" containerName="barbican-db-sync" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.232291 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.234608 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.234920 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2fh5m" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.238118 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.255462 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5544d45d85-dqp9x"] Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.282088 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6bc6797fd8-gflqx"] Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.285763 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.289687 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.327957 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6bc6797fd8-gflqx"] Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.352850 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffd8547bc-kfdwq"] Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.386904 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cddb74997-pr26l"] Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.389550 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.400348 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-combined-ca-bundle\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.400894 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2vdj\" (UniqueName: \"kubernetes.io/projected/f9973055-9eec-4599-9d2b-ccef87f6a481-kube-api-access-s2vdj\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.400919 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-combined-ca-bundle\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.400959 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9973055-9eec-4599-9d2b-ccef87f6a481-logs\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.400978 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.402091 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data-custom\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.402147 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8724dd55-b509-4056-b324-531ec57c3a7e-logs\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.402228 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data-custom\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.402249 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9sj8\" (UniqueName: \"kubernetes.io/projected/8724dd55-b509-4056-b324-531ec57c3a7e-kube-api-access-h9sj8\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.402455 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.445076 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-pr26l"] Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.466789 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-64f44fdcb4-24cj8"] Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.472857 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.489502 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.500382 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64f44fdcb4-24cj8"] Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.506821 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8724dd55-b509-4056-b324-531ec57c3a7e-logs\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510276 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8724dd55-b509-4056-b324-531ec57c3a7e-logs\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510379 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data-custom\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510414 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9sj8\" (UniqueName: \"kubernetes.io/projected/8724dd55-b509-4056-b324-531ec57c3a7e-kube-api-access-h9sj8\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510486 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-svc\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510521 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8449c\" (UniqueName: \"kubernetes.io/projected/8fa26949-291d-4703-b9f0-b31858214f37-kube-api-access-8449c\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510557 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-swift-storage-0\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510587 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-nb\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510638 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510668 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-sb\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510723 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-combined-ca-bundle\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510769 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2vdj\" (UniqueName: \"kubernetes.io/projected/f9973055-9eec-4599-9d2b-ccef87f6a481-kube-api-access-s2vdj\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510812 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-combined-ca-bundle\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510856 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9973055-9eec-4599-9d2b-ccef87f6a481-logs\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510900 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510932 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-config\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.510964 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data-custom\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.518557 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9973055-9eec-4599-9d2b-ccef87f6a481-logs\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.526222 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.537450 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.545831 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-combined-ca-bundle\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.560166 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data-custom\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.581912 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-combined-ca-bundle\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.599415 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2vdj\" (UniqueName: \"kubernetes.io/projected/f9973055-9eec-4599-9d2b-ccef87f6a481-kube-api-access-s2vdj\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.599605 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data-custom\") pod \"barbican-worker-5544d45d85-dqp9x\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.600464 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9sj8\" (UniqueName: \"kubernetes.io/projected/8724dd55-b509-4056-b324-531ec57c3a7e-kube-api-access-h9sj8\") pod \"barbican-keystone-listener-6bc6797fd8-gflqx\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.616937 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-sb\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.618391 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/998a283e-c92c-4707-af67-a200119a8d75-logs\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.618559 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-combined-ca-bundle\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.618664 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-config\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.618697 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skjxv\" (UniqueName: \"kubernetes.io/projected/998a283e-c92c-4707-af67-a200119a8d75-kube-api-access-skjxv\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.618867 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data-custom\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.618904 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-svc\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.618921 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.618967 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8449c\" (UniqueName: \"kubernetes.io/projected/8fa26949-291d-4703-b9f0-b31858214f37-kube-api-access-8449c\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.619020 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-swift-storage-0\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.619060 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-nb\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.620007 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-nb\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.620526 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-sb\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.623124 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.624974 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-config\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.625136 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-svc\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.625223 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-swift-storage-0\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.643625 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8449c\" (UniqueName: \"kubernetes.io/projected/8fa26949-291d-4703-b9f0-b31858214f37-kube-api-access-8449c\") pod \"dnsmasq-dns-cddb74997-pr26l\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.720597 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-combined-ca-bundle\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.720681 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skjxv\" (UniqueName: \"kubernetes.io/projected/998a283e-c92c-4707-af67-a200119a8d75-kube-api-access-skjxv\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.720744 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data-custom\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.720768 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.720983 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/998a283e-c92c-4707-af67-a200119a8d75-logs\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.723646 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/998a283e-c92c-4707-af67-a200119a8d75-logs\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.727593 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data-custom\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.728251 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-combined-ca-bundle\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.729301 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.729844 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.748163 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skjxv\" (UniqueName: \"kubernetes.io/projected/998a283e-c92c-4707-af67-a200119a8d75-kube-api-access-skjxv\") pod \"barbican-api-64f44fdcb4-24cj8\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.827364 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.861594 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:29:33 crc kubenswrapper[4871]: I1007 22:29:33.953094 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6bc6797fd8-gflqx"] Oct 07 22:29:33 crc kubenswrapper[4871]: W1007 22:29:33.999268 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8724dd55_b509_4056_b324_531ec57c3a7e.slice/crio-8844a5d2ddf30eba5f81c5e76504bc6ea6fb435a7f5ff614b2673ba68c45ae7c WatchSource:0}: Error finding container 8844a5d2ddf30eba5f81c5e76504bc6ea6fb435a7f5ff614b2673ba68c45ae7c: Status 404 returned error can't find the container with id 8844a5d2ddf30eba5f81c5e76504bc6ea6fb435a7f5ff614b2673ba68c45ae7c Oct 07 22:29:34 crc kubenswrapper[4871]: I1007 22:29:34.337455 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-pr26l"] Oct 07 22:29:34 crc kubenswrapper[4871]: I1007 22:29:34.561014 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64f44fdcb4-24cj8"] Oct 07 22:29:34 crc kubenswrapper[4871]: W1007 22:29:34.571097 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod998a283e_c92c_4707_af67_a200119a8d75.slice/crio-888a858e3086aa8421042343f6ab37e66d90884e5142d991ac8d0021677d3f6a WatchSource:0}: Error finding container 888a858e3086aa8421042343f6ab37e66d90884e5142d991ac8d0021677d3f6a: Status 404 returned error can't find the container with id 888a858e3086aa8421042343f6ab37e66d90884e5142d991ac8d0021677d3f6a Oct 07 22:29:34 crc kubenswrapper[4871]: W1007 22:29:34.662715 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9973055_9eec_4599_9d2b_ccef87f6a481.slice/crio-f9ba6f90866aa6a695fac71c4be8863f248545b55cda441d59d486660156a5b0 WatchSource:0}: Error finding container f9ba6f90866aa6a695fac71c4be8863f248545b55cda441d59d486660156a5b0: Status 404 returned error can't find the container with id f9ba6f90866aa6a695fac71c4be8863f248545b55cda441d59d486660156a5b0 Oct 07 22:29:34 crc kubenswrapper[4871]: I1007 22:29:34.664619 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5544d45d85-dqp9x"] Oct 07 22:29:34 crc kubenswrapper[4871]: I1007 22:29:34.888203 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64f44fdcb4-24cj8" event={"ID":"998a283e-c92c-4707-af67-a200119a8d75","Type":"ContainerStarted","Data":"888a858e3086aa8421042343f6ab37e66d90884e5142d991ac8d0021677d3f6a"} Oct 07 22:29:34 crc kubenswrapper[4871]: I1007 22:29:34.892767 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5544d45d85-dqp9x" event={"ID":"f9973055-9eec-4599-9d2b-ccef87f6a481","Type":"ContainerStarted","Data":"f9ba6f90866aa6a695fac71c4be8863f248545b55cda441d59d486660156a5b0"} Oct 07 22:29:34 crc kubenswrapper[4871]: I1007 22:29:34.894119 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-pr26l" event={"ID":"8fa26949-291d-4703-b9f0-b31858214f37","Type":"ContainerStarted","Data":"9c354eb3929e4aeb8bf48ab17f0f821f441bd140d110912197afa96c9f9c1d65"} Oct 07 22:29:34 crc kubenswrapper[4871]: I1007 22:29:34.895498 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" event={"ID":"8724dd55-b509-4056-b324-531ec57c3a7e","Type":"ContainerStarted","Data":"8844a5d2ddf30eba5f81c5e76504bc6ea6fb435a7f5ff614b2673ba68c45ae7c"} Oct 07 22:29:34 crc kubenswrapper[4871]: I1007 22:29:34.895729 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" podUID="0cdd7757-e120-4571-87e8-85538c0c50b0" containerName="dnsmasq-dns" containerID="cri-o://3e1eaadc61aed052ef0e52fa6bd17cd42ed8c865d84e7d580bf6f63987ba92fe" gracePeriod=10 Oct 07 22:29:35 crc kubenswrapper[4871]: I1007 22:29:35.921904 4871 generic.go:334] "Generic (PLEG): container finished" podID="8fa26949-291d-4703-b9f0-b31858214f37" containerID="0b6868958e3db44b4eb5bfa5d346792a5289cb732b07bd70d215cd1b8ea38450" exitCode=0 Oct 07 22:29:35 crc kubenswrapper[4871]: I1007 22:29:35.922844 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-pr26l" event={"ID":"8fa26949-291d-4703-b9f0-b31858214f37","Type":"ContainerDied","Data":"0b6868958e3db44b4eb5bfa5d346792a5289cb732b07bd70d215cd1b8ea38450"} Oct 07 22:29:35 crc kubenswrapper[4871]: I1007 22:29:35.941639 4871 generic.go:334] "Generic (PLEG): container finished" podID="0cdd7757-e120-4571-87e8-85538c0c50b0" containerID="3e1eaadc61aed052ef0e52fa6bd17cd42ed8c865d84e7d580bf6f63987ba92fe" exitCode=0 Oct 07 22:29:35 crc kubenswrapper[4871]: I1007 22:29:35.941827 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" event={"ID":"0cdd7757-e120-4571-87e8-85538c0c50b0","Type":"ContainerDied","Data":"3e1eaadc61aed052ef0e52fa6bd17cd42ed8c865d84e7d580bf6f63987ba92fe"} Oct 07 22:29:35 crc kubenswrapper[4871]: I1007 22:29:35.948051 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cc4f647bc-9lrsh" event={"ID":"7116e0fa-30d9-4a8d-9a11-0d57908f268f","Type":"ContainerStarted","Data":"09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0"} Oct 07 22:29:35 crc kubenswrapper[4871]: I1007 22:29:35.954335 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64f44fdcb4-24cj8" event={"ID":"998a283e-c92c-4707-af67-a200119a8d75","Type":"ContainerStarted","Data":"747581504c4cb242ca1b5b8ac1ab72c6f8680a6a27a5e1f89f29c995e71d07fa"} Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.786967 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-677599d776-jjqd2"] Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.789198 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.792222 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-677599d776-jjqd2"] Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.802156 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.806353 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.808857 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-public-tls-certs\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.808928 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c750808e-de6d-442b-82ed-d3498b036d91-logs\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.809001 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-internal-tls-certs\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.809100 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data-custom\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.809140 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-combined-ca-bundle\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.809169 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.809349 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96hnw\" (UniqueName: \"kubernetes.io/projected/c750808e-de6d-442b-82ed-d3498b036d91-kube-api-access-96hnw\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.912187 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c750808e-de6d-442b-82ed-d3498b036d91-logs\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.912279 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-internal-tls-certs\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.912323 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data-custom\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.912359 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-combined-ca-bundle\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.912383 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.912432 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96hnw\" (UniqueName: \"kubernetes.io/projected/c750808e-de6d-442b-82ed-d3498b036d91-kube-api-access-96hnw\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.912953 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c750808e-de6d-442b-82ed-d3498b036d91-logs\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.913730 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-public-tls-certs\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.922894 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-combined-ca-bundle\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.928379 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-public-tls-certs\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.929629 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.932497 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-internal-tls-certs\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.933834 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96hnw\" (UniqueName: \"kubernetes.io/projected/c750808e-de6d-442b-82ed-d3498b036d91-kube-api-access-96hnw\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:36 crc kubenswrapper[4871]: I1007 22:29:36.942718 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data-custom\") pod \"barbican-api-677599d776-jjqd2\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:37 crc kubenswrapper[4871]: I1007 22:29:37.119358 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.222933 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.377431 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-sb\") pod \"0cdd7757-e120-4571-87e8-85538c0c50b0\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.377931 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-svc\") pod \"0cdd7757-e120-4571-87e8-85538c0c50b0\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.378179 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxdr5\" (UniqueName: \"kubernetes.io/projected/0cdd7757-e120-4571-87e8-85538c0c50b0-kube-api-access-vxdr5\") pod \"0cdd7757-e120-4571-87e8-85538c0c50b0\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.378231 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-swift-storage-0\") pod \"0cdd7757-e120-4571-87e8-85538c0c50b0\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.378277 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-config\") pod \"0cdd7757-e120-4571-87e8-85538c0c50b0\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.378301 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-nb\") pod \"0cdd7757-e120-4571-87e8-85538c0c50b0\" (UID: \"0cdd7757-e120-4571-87e8-85538c0c50b0\") " Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.387015 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cdd7757-e120-4571-87e8-85538c0c50b0-kube-api-access-vxdr5" (OuterVolumeSpecName: "kube-api-access-vxdr5") pod "0cdd7757-e120-4571-87e8-85538c0c50b0" (UID: "0cdd7757-e120-4571-87e8-85538c0c50b0"). InnerVolumeSpecName "kube-api-access-vxdr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.440415 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0cdd7757-e120-4571-87e8-85538c0c50b0" (UID: "0cdd7757-e120-4571-87e8-85538c0c50b0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.442760 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0cdd7757-e120-4571-87e8-85538c0c50b0" (UID: "0cdd7757-e120-4571-87e8-85538c0c50b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.449660 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0cdd7757-e120-4571-87e8-85538c0c50b0" (UID: "0cdd7757-e120-4571-87e8-85538c0c50b0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.459292 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0cdd7757-e120-4571-87e8-85538c0c50b0" (UID: "0cdd7757-e120-4571-87e8-85538c0c50b0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.459910 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-config" (OuterVolumeSpecName: "config") pod "0cdd7757-e120-4571-87e8-85538c0c50b0" (UID: "0cdd7757-e120-4571-87e8-85538c0c50b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.482604 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.482651 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.482662 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxdr5\" (UniqueName: \"kubernetes.io/projected/0cdd7757-e120-4571-87e8-85538c0c50b0-kube-api-access-vxdr5\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.482674 4871 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.482683 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:39 crc kubenswrapper[4871]: I1007 22:29:39.482691 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cdd7757-e120-4571-87e8-85538c0c50b0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:40 crc kubenswrapper[4871]: I1007 22:29:40.003582 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" event={"ID":"0cdd7757-e120-4571-87e8-85538c0c50b0","Type":"ContainerDied","Data":"7984a611b854beaa53e0e794d0e0265c16c4a418efe40723276b6a387e6402d5"} Oct 07 22:29:40 crc kubenswrapper[4871]: I1007 22:29:40.003642 4871 scope.go:117] "RemoveContainer" containerID="3e1eaadc61aed052ef0e52fa6bd17cd42ed8c865d84e7d580bf6f63987ba92fe" Oct 07 22:29:40 crc kubenswrapper[4871]: I1007 22:29:40.003779 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffd8547bc-kfdwq" Oct 07 22:29:40 crc kubenswrapper[4871]: I1007 22:29:40.045666 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffd8547bc-kfdwq"] Oct 07 22:29:40 crc kubenswrapper[4871]: I1007 22:29:40.053242 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6ffd8547bc-kfdwq"] Oct 07 22:29:40 crc kubenswrapper[4871]: I1007 22:29:40.501819 4871 scope.go:117] "RemoveContainer" containerID="c034333d104008806b90644a8cd76a24d9025f0b6ef3580a98e3110ee7931157" Oct 07 22:29:40 crc kubenswrapper[4871]: I1007 22:29:40.899856 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-677599d776-jjqd2"] Oct 07 22:29:40 crc kubenswrapper[4871]: I1007 22:29:40.997882 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cdd7757-e120-4571-87e8-85538c0c50b0" path="/var/lib/kubelet/pods/0cdd7757-e120-4571-87e8-85538c0c50b0/volumes" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.038698 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" event={"ID":"8724dd55-b509-4056-b324-531ec57c3a7e","Type":"ContainerStarted","Data":"d4cf3f49a66e3bfb1a2f0116e08e29835a98701bf050feff3371d723047d7b5c"} Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.044923 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677599d776-jjqd2" event={"ID":"c750808e-de6d-442b-82ed-d3498b036d91","Type":"ContainerStarted","Data":"3ce17290b46def925cf8b9b69d4142ed8a8f644537f24587c21309531f91602e"} Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.044989 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677599d776-jjqd2" event={"ID":"c750808e-de6d-442b-82ed-d3498b036d91","Type":"ContainerStarted","Data":"94b95aca5eb11093877243f356cbb8de96ca88027ee3aa0bdfdcca7dab2af178"} Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.050820 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64f44fdcb4-24cj8" event={"ID":"998a283e-c92c-4707-af67-a200119a8d75","Type":"ContainerStarted","Data":"39478530ec510d4ac29858ac7d784223b49f3d347d35c1d00f549244862145e4"} Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.051502 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.051851 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.066020 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5544d45d85-dqp9x" event={"ID":"f9973055-9eec-4599-9d2b-ccef87f6a481","Type":"ContainerStarted","Data":"26e20164e3525af55f84ab8ba8cc88a36fbcc0795cd362085e6b25b516d314d7"} Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.072409 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-64f44fdcb4-24cj8" podStartSLOduration=9.072385177 podStartE2EDuration="9.072385177s" podCreationTimestamp="2025-10-07 22:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:42.071519994 +0000 UTC m=+1255.874218067" watchObservedRunningTime="2025-10-07 22:29:42.072385177 +0000 UTC m=+1255.875083250" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.079696 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-pr26l" event={"ID":"8fa26949-291d-4703-b9f0-b31858214f37","Type":"ContainerStarted","Data":"e48e3b09bb6b842b3e6529b901cd5fdd225e0c407aca4b30c7dce97f207717ab"} Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.080046 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.098831 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerStarted","Data":"528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420"} Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.099023 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="ceilometer-central-agent" containerID="cri-o://eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46" gracePeriod=30 Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.099254 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.099303 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="proxy-httpd" containerID="cri-o://528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420" gracePeriod=30 Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.099351 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="sg-core" containerID="cri-o://b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38" gracePeriod=30 Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.099394 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="ceilometer-notification-agent" containerID="cri-o://0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed" gracePeriod=30 Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.110098 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cc4f647bc-9lrsh" event={"ID":"7116e0fa-30d9-4a8d-9a11-0d57908f268f","Type":"ContainerStarted","Data":"48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a"} Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.110462 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.122074 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5544d45d85-dqp9x" podStartSLOduration=2.440847136 podStartE2EDuration="9.122045524s" podCreationTimestamp="2025-10-07 22:29:33 +0000 UTC" firstStartedPulling="2025-10-07 22:29:34.666061354 +0000 UTC m=+1248.468759427" lastFinishedPulling="2025-10-07 22:29:41.347259742 +0000 UTC m=+1255.149957815" observedRunningTime="2025-10-07 22:29:42.098038827 +0000 UTC m=+1255.900736900" watchObservedRunningTime="2025-10-07 22:29:42.122045524 +0000 UTC m=+1255.924743597" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.145035 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cddb74997-pr26l" podStartSLOduration=9.145005512000001 podStartE2EDuration="9.145005512s" podCreationTimestamp="2025-10-07 22:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:42.13434338 +0000 UTC m=+1255.937041453" watchObservedRunningTime="2025-10-07 22:29:42.145005512 +0000 UTC m=+1255.947703585" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.179722 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7cc4f647bc-9lrsh" podStartSLOduration=11.179692812 podStartE2EDuration="11.179692812s" podCreationTimestamp="2025-10-07 22:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:42.159810285 +0000 UTC m=+1255.962508358" watchObservedRunningTime="2025-10-07 22:29:42.179692812 +0000 UTC m=+1255.982390895" Oct 07 22:29:42 crc kubenswrapper[4871]: I1007 22:29:42.195597 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.925101924 podStartE2EDuration="49.195568883s" podCreationTimestamp="2025-10-07 22:28:53 +0000 UTC" firstStartedPulling="2025-10-07 22:28:54.071846892 +0000 UTC m=+1207.874544965" lastFinishedPulling="2025-10-07 22:29:41.342313851 +0000 UTC m=+1255.145011924" observedRunningTime="2025-10-07 22:29:42.1845168 +0000 UTC m=+1255.987214873" watchObservedRunningTime="2025-10-07 22:29:42.195568883 +0000 UTC m=+1255.998266956" Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.121552 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" event={"ID":"8724dd55-b509-4056-b324-531ec57c3a7e","Type":"ContainerStarted","Data":"2a05385f6943057613877aa85b019b47371856ca80b0e282c3f1b443a5093210"} Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.124642 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7sq86" event={"ID":"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6","Type":"ContainerStarted","Data":"5b46a03277275d9c2cc38bd8814c0a9e068549a5c9ce5f29fba665667d373c7a"} Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.128341 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5544d45d85-dqp9x" event={"ID":"f9973055-9eec-4599-9d2b-ccef87f6a481","Type":"ContainerStarted","Data":"f1bb4a9d4f25de043a799d48cedbb3a8b9bba20dd97ad863253a098ddb8d0065"} Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.132482 4871 generic.go:334] "Generic (PLEG): container finished" podID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerID="528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420" exitCode=0 Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.132517 4871 generic.go:334] "Generic (PLEG): container finished" podID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerID="b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38" exitCode=2 Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.132533 4871 generic.go:334] "Generic (PLEG): container finished" podID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerID="eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46" exitCode=0 Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.132591 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerDied","Data":"528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420"} Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.132853 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerDied","Data":"b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38"} Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.132878 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerDied","Data":"eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46"} Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.137148 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677599d776-jjqd2" event={"ID":"c750808e-de6d-442b-82ed-d3498b036d91","Type":"ContainerStarted","Data":"e6f0d276a1846b36e45c6f4882a0d1c7b6023b4bc31e6f161d4fcc4d02c212f3"} Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.137218 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.138914 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.154527 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" podStartSLOduration=2.808362911 podStartE2EDuration="10.154504128s" podCreationTimestamp="2025-10-07 22:29:33 +0000 UTC" firstStartedPulling="2025-10-07 22:29:34.002116851 +0000 UTC m=+1247.804814924" lastFinishedPulling="2025-10-07 22:29:41.348258068 +0000 UTC m=+1255.150956141" observedRunningTime="2025-10-07 22:29:43.148726504 +0000 UTC m=+1256.951424617" watchObservedRunningTime="2025-10-07 22:29:43.154504128 +0000 UTC m=+1256.957202201" Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.192320 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7sq86" podStartSLOduration=7.736479418 podStartE2EDuration="44.192291309s" podCreationTimestamp="2025-10-07 22:28:59 +0000 UTC" firstStartedPulling="2025-10-07 22:29:04.891943794 +0000 UTC m=+1218.694641867" lastFinishedPulling="2025-10-07 22:29:41.347755665 +0000 UTC m=+1255.150453758" observedRunningTime="2025-10-07 22:29:43.16854677 +0000 UTC m=+1256.971244843" watchObservedRunningTime="2025-10-07 22:29:43.192291309 +0000 UTC m=+1256.994989392" Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.203708 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-677599d776-jjqd2" podStartSLOduration=7.203685472 podStartE2EDuration="7.203685472s" podCreationTimestamp="2025-10-07 22:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:43.189205948 +0000 UTC m=+1256.991904021" watchObservedRunningTime="2025-10-07 22:29:43.203685472 +0000 UTC m=+1257.006383545" Oct 07 22:29:43 crc kubenswrapper[4871]: I1007 22:29:43.844557 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.774953 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.915215 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-combined-ca-bundle\") pod \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.915312 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-scripts\") pod \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.915363 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-run-httpd\") pod \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.915440 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-sg-core-conf-yaml\") pod \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.915491 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-log-httpd\") pod \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.915528 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrldk\" (UniqueName: \"kubernetes.io/projected/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-kube-api-access-lrldk\") pod \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.915622 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-config-data\") pod \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\" (UID: \"2daf4b99-374d-45df-b28c-4f02cfa1dfa8\") " Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.915900 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2daf4b99-374d-45df-b28c-4f02cfa1dfa8" (UID: "2daf4b99-374d-45df-b28c-4f02cfa1dfa8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.916103 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.916170 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2daf4b99-374d-45df-b28c-4f02cfa1dfa8" (UID: "2daf4b99-374d-45df-b28c-4f02cfa1dfa8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.923901 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-scripts" (OuterVolumeSpecName: "scripts") pod "2daf4b99-374d-45df-b28c-4f02cfa1dfa8" (UID: "2daf4b99-374d-45df-b28c-4f02cfa1dfa8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.924045 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-kube-api-access-lrldk" (OuterVolumeSpecName: "kube-api-access-lrldk") pod "2daf4b99-374d-45df-b28c-4f02cfa1dfa8" (UID: "2daf4b99-374d-45df-b28c-4f02cfa1dfa8"). InnerVolumeSpecName "kube-api-access-lrldk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:44 crc kubenswrapper[4871]: I1007 22:29:44.949137 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2daf4b99-374d-45df-b28c-4f02cfa1dfa8" (UID: "2daf4b99-374d-45df-b28c-4f02cfa1dfa8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.003641 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2daf4b99-374d-45df-b28c-4f02cfa1dfa8" (UID: "2daf4b99-374d-45df-b28c-4f02cfa1dfa8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.018614 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.018658 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.018675 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.018688 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.018703 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrldk\" (UniqueName: \"kubernetes.io/projected/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-kube-api-access-lrldk\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.052808 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-config-data" (OuterVolumeSpecName: "config-data") pod "2daf4b99-374d-45df-b28c-4f02cfa1dfa8" (UID: "2daf4b99-374d-45df-b28c-4f02cfa1dfa8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.120949 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2daf4b99-374d-45df-b28c-4f02cfa1dfa8-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.170287 4871 generic.go:334] "Generic (PLEG): container finished" podID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerID="0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed" exitCode=0 Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.170395 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerDied","Data":"0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed"} Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.170440 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.170476 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2daf4b99-374d-45df-b28c-4f02cfa1dfa8","Type":"ContainerDied","Data":"d4b902047b20ab19d87d251e148df5e75c8fe64b09729d10ef1dd00e9d19234a"} Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.170498 4871 scope.go:117] "RemoveContainer" containerID="528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.204361 4871 scope.go:117] "RemoveContainer" containerID="b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.214598 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.222332 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.246485 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.247119 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="proxy-httpd" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247148 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="proxy-httpd" Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.247177 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="sg-core" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247196 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="sg-core" Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.247218 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cdd7757-e120-4571-87e8-85538c0c50b0" containerName="dnsmasq-dns" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247228 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cdd7757-e120-4571-87e8-85538c0c50b0" containerName="dnsmasq-dns" Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.247246 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cdd7757-e120-4571-87e8-85538c0c50b0" containerName="init" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247255 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cdd7757-e120-4571-87e8-85538c0c50b0" containerName="init" Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.247269 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="ceilometer-notification-agent" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247277 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="ceilometer-notification-agent" Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.247291 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="ceilometer-central-agent" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247299 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="ceilometer-central-agent" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247502 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cdd7757-e120-4571-87e8-85538c0c50b0" containerName="dnsmasq-dns" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247521 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="ceilometer-central-agent" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247546 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="proxy-httpd" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247559 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="sg-core" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247573 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" containerName="ceilometer-notification-agent" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.247171 4871 scope.go:117] "RemoveContainer" containerID="0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.271158 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.280920 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.281191 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.290287 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.321759 4871 scope.go:117] "RemoveContainer" containerID="eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.347104 4871 scope.go:117] "RemoveContainer" containerID="528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420" Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.347761 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420\": container with ID starting with 528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420 not found: ID does not exist" containerID="528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.347861 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420"} err="failed to get container status \"528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420\": rpc error: code = NotFound desc = could not find container \"528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420\": container with ID starting with 528b43bc625f07ae49124af0a190dd739aa775f2511e9e394892a4c928acc420 not found: ID does not exist" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.347923 4871 scope.go:117] "RemoveContainer" containerID="b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38" Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.348275 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38\": container with ID starting with b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38 not found: ID does not exist" containerID="b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.348325 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38"} err="failed to get container status \"b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38\": rpc error: code = NotFound desc = could not find container \"b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38\": container with ID starting with b9e3d976f2949acbcf7db310b4111c0c30ab0b2f1e6b9f2eece61b6be7eadb38 not found: ID does not exist" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.348342 4871 scope.go:117] "RemoveContainer" containerID="0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed" Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.348598 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed\": container with ID starting with 0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed not found: ID does not exist" containerID="0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.348640 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed"} err="failed to get container status \"0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed\": rpc error: code = NotFound desc = could not find container \"0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed\": container with ID starting with 0f761394340c4888ad39fa4a60fe41254be1ade6f501379d9ecd44d53f5925ed not found: ID does not exist" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.348656 4871 scope.go:117] "RemoveContainer" containerID="eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46" Oct 07 22:29:45 crc kubenswrapper[4871]: E1007 22:29:45.348915 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46\": container with ID starting with eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46 not found: ID does not exist" containerID="eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.349020 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46"} err="failed to get container status \"eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46\": rpc error: code = NotFound desc = could not find container \"eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46\": container with ID starting with eeb6047a2df164e24a32c883fb6a1558273ddc843011bdbdb4f5133283f7df46 not found: ID does not exist" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.429609 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-log-httpd\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.429733 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-run-httpd\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.429764 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.429966 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-config-data\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.430152 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v9cc\" (UniqueName: \"kubernetes.io/projected/ae96a676-3a8b-44f0-a54a-5031450388d1-kube-api-access-8v9cc\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.430197 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.430364 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-scripts\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.464708 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.479728 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.532586 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-scripts\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.533130 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-log-httpd\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.533193 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-run-httpd\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.533220 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.533263 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-config-data\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.533310 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v9cc\" (UniqueName: \"kubernetes.io/projected/ae96a676-3a8b-44f0-a54a-5031450388d1-kube-api-access-8v9cc\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.533363 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.533988 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-log-httpd\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.534240 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-run-httpd\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.544185 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.558920 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-scripts\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.559423 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.559979 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-config-data\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.566605 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v9cc\" (UniqueName: \"kubernetes.io/projected/ae96a676-3a8b-44f0-a54a-5031450388d1-kube-api-access-8v9cc\") pod \"ceilometer-0\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " pod="openstack/ceilometer-0" Oct 07 22:29:45 crc kubenswrapper[4871]: I1007 22:29:45.593275 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:29:46 crc kubenswrapper[4871]: I1007 22:29:46.116744 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:29:46 crc kubenswrapper[4871]: I1007 22:29:46.183681 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerStarted","Data":"81b8621a149a9b4fdc1c9d267bd6d1227c1fbf52f494d2b3c9fb498c983ab7fd"} Oct 07 22:29:47 crc kubenswrapper[4871]: I1007 22:29:47.003322 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2daf4b99-374d-45df-b28c-4f02cfa1dfa8" path="/var/lib/kubelet/pods/2daf4b99-374d-45df-b28c-4f02cfa1dfa8/volumes" Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.205621 4871 generic.go:334] "Generic (PLEG): container finished" podID="a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" containerID="5b46a03277275d9c2cc38bd8814c0a9e068549a5c9ce5f29fba665667d373c7a" exitCode=0 Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.205690 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7sq86" event={"ID":"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6","Type":"ContainerDied","Data":"5b46a03277275d9c2cc38bd8814c0a9e068549a5c9ce5f29fba665667d373c7a"} Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.210154 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerStarted","Data":"8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be"} Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.521696 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.635944 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.702366 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-64f44fdcb4-24cj8"] Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.702648 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-64f44fdcb4-24cj8" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api-log" containerID="cri-o://747581504c4cb242ca1b5b8ac1ab72c6f8680a6a27a5e1f89f29c995e71d07fa" gracePeriod=30 Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.704653 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-64f44fdcb4-24cj8" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api" containerID="cri-o://39478530ec510d4ac29858ac7d784223b49f3d347d35c1d00f549244862145e4" gracePeriod=30 Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.713724 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64f44fdcb4-24cj8" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": EOF" Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.732922 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.805389 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-k2t47"] Oct 07 22:29:48 crc kubenswrapper[4871]: I1007 22:29:48.805672 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" podUID="65670f97-1301-4482-bb5d-fc4464218f24" containerName="dnsmasq-dns" containerID="cri-o://d0390f1f00eaa4fb6e4dc40b42d55096049958ee99bfee6fbcb90c72fc6aac1a" gracePeriod=10 Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.234867 4871 generic.go:334] "Generic (PLEG): container finished" podID="65670f97-1301-4482-bb5d-fc4464218f24" containerID="d0390f1f00eaa4fb6e4dc40b42d55096049958ee99bfee6fbcb90c72fc6aac1a" exitCode=0 Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.235102 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" event={"ID":"65670f97-1301-4482-bb5d-fc4464218f24","Type":"ContainerDied","Data":"d0390f1f00eaa4fb6e4dc40b42d55096049958ee99bfee6fbcb90c72fc6aac1a"} Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.247223 4871 generic.go:334] "Generic (PLEG): container finished" podID="998a283e-c92c-4707-af67-a200119a8d75" containerID="747581504c4cb242ca1b5b8ac1ab72c6f8680a6a27a5e1f89f29c995e71d07fa" exitCode=143 Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.247968 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64f44fdcb4-24cj8" event={"ID":"998a283e-c92c-4707-af67-a200119a8d75","Type":"ContainerDied","Data":"747581504c4cb242ca1b5b8ac1ab72c6f8680a6a27a5e1f89f29c995e71d07fa"} Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.437397 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.532718 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-swift-storage-0\") pod \"65670f97-1301-4482-bb5d-fc4464218f24\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.532846 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-sb\") pod \"65670f97-1301-4482-bb5d-fc4464218f24\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.533026 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-svc\") pod \"65670f97-1301-4482-bb5d-fc4464218f24\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.533136 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krdjq\" (UniqueName: \"kubernetes.io/projected/65670f97-1301-4482-bb5d-fc4464218f24-kube-api-access-krdjq\") pod \"65670f97-1301-4482-bb5d-fc4464218f24\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.533196 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-config\") pod \"65670f97-1301-4482-bb5d-fc4464218f24\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.533218 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-nb\") pod \"65670f97-1301-4482-bb5d-fc4464218f24\" (UID: \"65670f97-1301-4482-bb5d-fc4464218f24\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.565818 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65670f97-1301-4482-bb5d-fc4464218f24-kube-api-access-krdjq" (OuterVolumeSpecName: "kube-api-access-krdjq") pod "65670f97-1301-4482-bb5d-fc4464218f24" (UID: "65670f97-1301-4482-bb5d-fc4464218f24"). InnerVolumeSpecName "kube-api-access-krdjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.636429 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krdjq\" (UniqueName: \"kubernetes.io/projected/65670f97-1301-4482-bb5d-fc4464218f24-kube-api-access-krdjq\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.663618 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "65670f97-1301-4482-bb5d-fc4464218f24" (UID: "65670f97-1301-4482-bb5d-fc4464218f24"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.669638 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-config" (OuterVolumeSpecName: "config") pod "65670f97-1301-4482-bb5d-fc4464218f24" (UID: "65670f97-1301-4482-bb5d-fc4464218f24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.673784 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "65670f97-1301-4482-bb5d-fc4464218f24" (UID: "65670f97-1301-4482-bb5d-fc4464218f24"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.698742 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "65670f97-1301-4482-bb5d-fc4464218f24" (UID: "65670f97-1301-4482-bb5d-fc4464218f24"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.699488 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7sq86" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.699739 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "65670f97-1301-4482-bb5d-fc4464218f24" (UID: "65670f97-1301-4482-bb5d-fc4464218f24"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.745495 4871 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.745541 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.745553 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.745564 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.745576 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65670f97-1301-4482-bb5d-fc4464218f24-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.851042 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-config-data\") pod \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.851083 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-etc-machine-id\") pod \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.851175 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-db-sync-config-data\") pod \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.851204 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-combined-ca-bundle\") pod \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.851232 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-scripts\") pod \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.851377 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hps9v\" (UniqueName: \"kubernetes.io/projected/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-kube-api-access-hps9v\") pod \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\" (UID: \"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6\") " Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.853951 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" (UID: "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.860934 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" (UID: "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.863945 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-kube-api-access-hps9v" (OuterVolumeSpecName: "kube-api-access-hps9v") pod "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" (UID: "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6"). InnerVolumeSpecName "kube-api-access-hps9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.867947 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-scripts" (OuterVolumeSpecName: "scripts") pod "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" (UID: "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.894944 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" (UID: "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.912998 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-config-data" (OuterVolumeSpecName: "config-data") pod "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" (UID: "a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.956489 4871 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.956521 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.956531 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.956543 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hps9v\" (UniqueName: \"kubernetes.io/projected/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-kube-api-access-hps9v\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.956557 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:49 crc kubenswrapper[4871]: I1007 22:29:49.956569 4871 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.266997 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" event={"ID":"65670f97-1301-4482-bb5d-fc4464218f24","Type":"ContainerDied","Data":"746a259f96ec23e61b97a16b0ba85c0ca471d0c11abd75dca9fdd979e5d11b0f"} Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.267524 4871 scope.go:117] "RemoveContainer" containerID="d0390f1f00eaa4fb6e4dc40b42d55096049958ee99bfee6fbcb90c72fc6aac1a" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.267019 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77dd5cf987-k2t47" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.271118 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7sq86" event={"ID":"a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6","Type":"ContainerDied","Data":"d74397ddaf67e27d6da2f67fa5ff091b2f2213bb519a7766c93af1c92ccb7aef"} Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.271167 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d74397ddaf67e27d6da2f67fa5ff091b2f2213bb519a7766c93af1c92ccb7aef" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.271237 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7sq86" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.297846 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerStarted","Data":"064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334"} Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.299365 4871 scope.go:117] "RemoveContainer" containerID="7752dee107e8a50cc0896cfb6745ee58b4e5272b92b6a05daf6722296ef556c7" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.354090 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-k2t47"] Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.390772 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.415581 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-k2t47"] Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.601883 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:29:50 crc kubenswrapper[4871]: E1007 22:29:50.602425 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65670f97-1301-4482-bb5d-fc4464218f24" containerName="init" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.602443 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="65670f97-1301-4482-bb5d-fc4464218f24" containerName="init" Oct 07 22:29:50 crc kubenswrapper[4871]: E1007 22:29:50.602474 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65670f97-1301-4482-bb5d-fc4464218f24" containerName="dnsmasq-dns" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.602482 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="65670f97-1301-4482-bb5d-fc4464218f24" containerName="dnsmasq-dns" Oct 07 22:29:50 crc kubenswrapper[4871]: E1007 22:29:50.602510 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" containerName="cinder-db-sync" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.602517 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" containerName="cinder-db-sync" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.602756 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" containerName="cinder-db-sync" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.602780 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="65670f97-1301-4482-bb5d-fc4464218f24" containerName="dnsmasq-dns" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.604013 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.610505 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.610859 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-lhvh2" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.611031 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.611138 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.623954 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.684701 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-thpsg"] Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.687237 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.711223 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.711307 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.711402 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhdd2\" (UniqueName: \"kubernetes.io/projected/5ec31882-e098-41e6-8c3f-f736f8092f9a-kube-api-access-zhdd2\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.711428 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec31882-e098-41e6-8c3f-f736f8092f9a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.711468 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-scripts\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.711488 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.713432 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-thpsg"] Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.795712 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.798843 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813067 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-sb\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813144 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhdd2\" (UniqueName: \"kubernetes.io/projected/5ec31882-e098-41e6-8c3f-f736f8092f9a-kube-api-access-zhdd2\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813170 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec31882-e098-41e6-8c3f-f736f8092f9a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813195 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-config\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813211 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxszr\" (UniqueName: \"kubernetes.io/projected/f73d961e-9fc8-48dd-baf3-56315f5b41a9-kube-api-access-jxszr\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813249 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-scripts\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813276 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813293 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-nb\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813331 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813374 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813414 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-svc\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.813439 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-swift-storage-0\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.814954 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec31882-e098-41e6-8c3f-f736f8092f9a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.827290 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.828151 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.828319 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.830430 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.838518 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-scripts\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.853213 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhdd2\" (UniqueName: \"kubernetes.io/projected/5ec31882-e098-41e6-8c3f-f736f8092f9a-kube-api-access-zhdd2\") pod \"cinder-scheduler-0\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " pod="openstack/cinder-scheduler-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.861406 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919009 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919128 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-svc\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919172 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-scripts\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919200 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-swift-storage-0\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919225 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79m2b\" (UniqueName: \"kubernetes.io/projected/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-kube-api-access-79m2b\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919282 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919319 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data-custom\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919345 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-sb\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919402 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-config\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919423 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxszr\" (UniqueName: \"kubernetes.io/projected/f73d961e-9fc8-48dd-baf3-56315f5b41a9-kube-api-access-jxszr\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919461 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-logs\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919506 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-nb\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.919532 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.920775 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-svc\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.921044 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-swift-storage-0\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.921184 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-sb\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.921579 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-nb\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.921863 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-config\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.948196 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxszr\" (UniqueName: \"kubernetes.io/projected/f73d961e-9fc8-48dd-baf3-56315f5b41a9-kube-api-access-jxszr\") pod \"dnsmasq-dns-59b9656b65-thpsg\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:50 crc kubenswrapper[4871]: I1007 22:29:50.961345 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.010409 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65670f97-1301-4482-bb5d-fc4464218f24" path="/var/lib/kubelet/pods/65670f97-1301-4482-bb5d-fc4464218f24/volumes" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.023466 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.024893 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.025625 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.026559 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-scripts\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.026606 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79m2b\" (UniqueName: \"kubernetes.io/projected/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-kube-api-access-79m2b\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.026918 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.027006 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data-custom\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.028316 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.028812 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-logs\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.029169 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-logs\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.033836 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data-custom\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.049803 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-scripts\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.052301 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79m2b\" (UniqueName: \"kubernetes.io/projected/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-kube-api-access-79m2b\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.054496 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.059156 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.128723 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.601415 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.795490 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-thpsg"] Oct 07 22:29:51 crc kubenswrapper[4871]: I1007 22:29:51.951172 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:29:52 crc kubenswrapper[4871]: I1007 22:29:52.349443 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ec31882-e098-41e6-8c3f-f736f8092f9a","Type":"ContainerStarted","Data":"52751a3dc502aa5cce24e5058c38e2cd54e853850c6485d21e13ab4ff5bec8f7"} Oct 07 22:29:52 crc kubenswrapper[4871]: I1007 22:29:52.352182 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4","Type":"ContainerStarted","Data":"8772f2b94b576d9a81ae95b313bd597ae0e324b3832b463300a86366969d89c5"} Oct 07 22:29:52 crc kubenswrapper[4871]: I1007 22:29:52.362372 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerStarted","Data":"6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d"} Oct 07 22:29:52 crc kubenswrapper[4871]: I1007 22:29:52.371390 4871 generic.go:334] "Generic (PLEG): container finished" podID="f73d961e-9fc8-48dd-baf3-56315f5b41a9" containerID="7688829b56e07f7ed2f0e0dc39b28809ded0a114a36e436cd45d912e15025bf5" exitCode=0 Oct 07 22:29:52 crc kubenswrapper[4871]: I1007 22:29:52.371447 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" event={"ID":"f73d961e-9fc8-48dd-baf3-56315f5b41a9","Type":"ContainerDied","Data":"7688829b56e07f7ed2f0e0dc39b28809ded0a114a36e436cd45d912e15025bf5"} Oct 07 22:29:52 crc kubenswrapper[4871]: I1007 22:29:52.371482 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" event={"ID":"f73d961e-9fc8-48dd-baf3-56315f5b41a9","Type":"ContainerStarted","Data":"b2264180a1a26dd81bd884254bfa3c483873211f6bf0e94661a36bde5528a95a"} Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.282124 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.326490 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.328305 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.335510 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.335825 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.336030 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-b95l4" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.337603 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.424908 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-openstack-config-secret\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.424990 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/943e025c-54d8-4686-81ee-894259d7af56-openstack-config\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.425037 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-combined-ca-bundle\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.425150 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b245b\" (UniqueName: \"kubernetes.io/projected/943e025c-54d8-4686-81ee-894259d7af56-kube-api-access-b245b\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.429596 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4","Type":"ContainerStarted","Data":"c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e"} Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.446291 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" event={"ID":"f73d961e-9fc8-48dd-baf3-56315f5b41a9","Type":"ContainerStarted","Data":"99727673524e91dec44e277b7293baff9da10c7f31708a6b980163103ba5f999"} Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.447544 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.499603 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" podStartSLOduration=3.499573136 podStartE2EDuration="3.499573136s" podCreationTimestamp="2025-10-07 22:29:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:53.490387603 +0000 UTC m=+1267.293085676" watchObservedRunningTime="2025-10-07 22:29:53.499573136 +0000 UTC m=+1267.302271209" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.527475 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-openstack-config-secret\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.527545 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/943e025c-54d8-4686-81ee-894259d7af56-openstack-config\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.529774 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/943e025c-54d8-4686-81ee-894259d7af56-openstack-config\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.529957 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-combined-ca-bundle\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.530249 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b245b\" (UniqueName: \"kubernetes.io/projected/943e025c-54d8-4686-81ee-894259d7af56-kube-api-access-b245b\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.533304 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-openstack-config-secret\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.541012 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-combined-ca-bundle\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.550015 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b245b\" (UniqueName: \"kubernetes.io/projected/943e025c-54d8-4686-81ee-894259d7af56-kube-api-access-b245b\") pod \"openstackclient\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.680953 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.928137 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64f44fdcb4-24cj8" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 22:29:53 crc kubenswrapper[4871]: I1007 22:29:53.928598 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64f44fdcb4-24cj8" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.145649 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64f44fdcb4-24cj8" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:36056->10.217.0.157:9311: read: connection reset by peer" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.145672 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64f44fdcb4-24cj8" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:36068->10.217.0.157:9311: read: connection reset by peer" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.194981 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.466852 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ec31882-e098-41e6-8c3f-f736f8092f9a","Type":"ContainerStarted","Data":"c5d24655de275a06d27468ec9a99b8b3dbac30cf3e541319a66a8b4b54758f3e"} Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.478632 4871 generic.go:334] "Generic (PLEG): container finished" podID="998a283e-c92c-4707-af67-a200119a8d75" containerID="39478530ec510d4ac29858ac7d784223b49f3d347d35c1d00f549244862145e4" exitCode=0 Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.478680 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64f44fdcb4-24cj8" event={"ID":"998a283e-c92c-4707-af67-a200119a8d75","Type":"ContainerDied","Data":"39478530ec510d4ac29858ac7d784223b49f3d347d35c1d00f549244862145e4"} Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.479611 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"943e025c-54d8-4686-81ee-894259d7af56","Type":"ContainerStarted","Data":"244ee5bbb1d61e2b5ca45ba763a89eb4eb3eaf2e7d0c2f9acdb17942291037a2"} Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.497867 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4","Type":"ContainerStarted","Data":"4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22"} Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.498075 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerName="cinder-api-log" containerID="cri-o://c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e" gracePeriod=30 Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.498186 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.498805 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerName="cinder-api" containerID="cri-o://4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22" gracePeriod=30 Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.537440 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.537415692 podStartE2EDuration="4.537415692s" podCreationTimestamp="2025-10-07 22:29:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:54.531177876 +0000 UTC m=+1268.333875959" watchObservedRunningTime="2025-10-07 22:29:54.537415692 +0000 UTC m=+1268.340113765" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.683674 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.777597 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data\") pod \"998a283e-c92c-4707-af67-a200119a8d75\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.777761 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-combined-ca-bundle\") pod \"998a283e-c92c-4707-af67-a200119a8d75\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.777821 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/998a283e-c92c-4707-af67-a200119a8d75-logs\") pod \"998a283e-c92c-4707-af67-a200119a8d75\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.777918 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data-custom\") pod \"998a283e-c92c-4707-af67-a200119a8d75\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.778108 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skjxv\" (UniqueName: \"kubernetes.io/projected/998a283e-c92c-4707-af67-a200119a8d75-kube-api-access-skjxv\") pod \"998a283e-c92c-4707-af67-a200119a8d75\" (UID: \"998a283e-c92c-4707-af67-a200119a8d75\") " Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.779022 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/998a283e-c92c-4707-af67-a200119a8d75-logs" (OuterVolumeSpecName: "logs") pod "998a283e-c92c-4707-af67-a200119a8d75" (UID: "998a283e-c92c-4707-af67-a200119a8d75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.795580 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/998a283e-c92c-4707-af67-a200119a8d75-kube-api-access-skjxv" (OuterVolumeSpecName: "kube-api-access-skjxv") pod "998a283e-c92c-4707-af67-a200119a8d75" (UID: "998a283e-c92c-4707-af67-a200119a8d75"). InnerVolumeSpecName "kube-api-access-skjxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.806103 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "998a283e-c92c-4707-af67-a200119a8d75" (UID: "998a283e-c92c-4707-af67-a200119a8d75"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.807966 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "998a283e-c92c-4707-af67-a200119a8d75" (UID: "998a283e-c92c-4707-af67-a200119a8d75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.861101 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data" (OuterVolumeSpecName: "config-data") pod "998a283e-c92c-4707-af67-a200119a8d75" (UID: "998a283e-c92c-4707-af67-a200119a8d75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.880439 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skjxv\" (UniqueName: \"kubernetes.io/projected/998a283e-c92c-4707-af67-a200119a8d75-kube-api-access-skjxv\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.880479 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.880490 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.880501 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/998a283e-c92c-4707-af67-a200119a8d75-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:54 crc kubenswrapper[4871]: I1007 22:29:54.880510 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/998a283e-c92c-4707-af67-a200119a8d75-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.315211 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.491854 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79m2b\" (UniqueName: \"kubernetes.io/projected/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-kube-api-access-79m2b\") pod \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.491909 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-etc-machine-id\") pod \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.492005 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-logs\") pod \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.492055 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data-custom\") pod \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.492081 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" (UID: "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.492162 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data\") pod \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.492197 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-scripts\") pod \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.492249 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-combined-ca-bundle\") pod \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\" (UID: \"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4\") " Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.492551 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-logs" (OuterVolumeSpecName: "logs") pod "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" (UID: "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.493269 4871 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.493293 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.500174 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-scripts" (OuterVolumeSpecName: "scripts") pod "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" (UID: "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.506736 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" (UID: "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.518234 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-kube-api-access-79m2b" (OuterVolumeSpecName: "kube-api-access-79m2b") pod "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" (UID: "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4"). InnerVolumeSpecName "kube-api-access-79m2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.523184 4871 generic.go:334] "Generic (PLEG): container finished" podID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerID="4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22" exitCode=0 Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.523229 4871 generic.go:334] "Generic (PLEG): container finished" podID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerID="c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e" exitCode=143 Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.523287 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4","Type":"ContainerDied","Data":"4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22"} Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.523371 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4","Type":"ContainerDied","Data":"c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e"} Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.523385 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4","Type":"ContainerDied","Data":"8772f2b94b576d9a81ae95b313bd597ae0e324b3832b463300a86366969d89c5"} Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.523402 4871 scope.go:117] "RemoveContainer" containerID="4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.523572 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.534345 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerStarted","Data":"5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40"} Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.535944 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.542325 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ec31882-e098-41e6-8c3f-f736f8092f9a","Type":"ContainerStarted","Data":"48ac3b9150034988c91416bc40d56e8a227416f00586e175b4460e93eb981b0f"} Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.557172 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" (UID: "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.566836 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.389662884 podStartE2EDuration="10.566815894s" podCreationTimestamp="2025-10-07 22:29:45 +0000 UTC" firstStartedPulling="2025-10-07 22:29:46.125436086 +0000 UTC m=+1259.928134159" lastFinishedPulling="2025-10-07 22:29:54.302589096 +0000 UTC m=+1268.105287169" observedRunningTime="2025-10-07 22:29:55.55910388 +0000 UTC m=+1269.361801943" watchObservedRunningTime="2025-10-07 22:29:55.566815894 +0000 UTC m=+1269.369513967" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.572691 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64f44fdcb4-24cj8" event={"ID":"998a283e-c92c-4707-af67-a200119a8d75","Type":"ContainerDied","Data":"888a858e3086aa8421042343f6ab37e66d90884e5142d991ac8d0021677d3f6a"} Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.572809 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64f44fdcb4-24cj8" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.590882 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.75889185 podStartE2EDuration="5.590858132s" podCreationTimestamp="2025-10-07 22:29:50 +0000 UTC" firstStartedPulling="2025-10-07 22:29:51.611958859 +0000 UTC m=+1265.414656932" lastFinishedPulling="2025-10-07 22:29:53.443925141 +0000 UTC m=+1267.246623214" observedRunningTime="2025-10-07 22:29:55.586215649 +0000 UTC m=+1269.388913722" watchObservedRunningTime="2025-10-07 22:29:55.590858132 +0000 UTC m=+1269.393556195" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.590966 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data" (OuterVolumeSpecName: "config-data") pod "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" (UID: "3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.596303 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79m2b\" (UniqueName: \"kubernetes.io/projected/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-kube-api-access-79m2b\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.596437 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.596505 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.596561 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.596615 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.599121 4871 scope.go:117] "RemoveContainer" containerID="c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.627942 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-64f44fdcb4-24cj8"] Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.638195 4871 scope.go:117] "RemoveContainer" containerID="4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22" Oct 07 22:29:55 crc kubenswrapper[4871]: E1007 22:29:55.641179 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22\": container with ID starting with 4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22 not found: ID does not exist" containerID="4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.641241 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22"} err="failed to get container status \"4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22\": rpc error: code = NotFound desc = could not find container \"4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22\": container with ID starting with 4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22 not found: ID does not exist" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.641272 4871 scope.go:117] "RemoveContainer" containerID="c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e" Oct 07 22:29:55 crc kubenswrapper[4871]: E1007 22:29:55.643349 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e\": container with ID starting with c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e not found: ID does not exist" containerID="c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.643614 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e"} err="failed to get container status \"c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e\": rpc error: code = NotFound desc = could not find container \"c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e\": container with ID starting with c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e not found: ID does not exist" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.643722 4871 scope.go:117] "RemoveContainer" containerID="4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.644371 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22"} err="failed to get container status \"4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22\": rpc error: code = NotFound desc = could not find container \"4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22\": container with ID starting with 4421ecde2aef00f9ca0b15c2b1d0839d5b4116b816239836d32ef713c9641a22 not found: ID does not exist" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.644465 4871 scope.go:117] "RemoveContainer" containerID="c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.644978 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e"} err="failed to get container status \"c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e\": rpc error: code = NotFound desc = could not find container \"c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e\": container with ID starting with c9599feef4b34d72a770ce021a653c0e0175abed02f0ab4914b1bfb9af61c76e not found: ID does not exist" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.645048 4871 scope.go:117] "RemoveContainer" containerID="39478530ec510d4ac29858ac7d784223b49f3d347d35c1d00f549244862145e4" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.645034 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-64f44fdcb4-24cj8"] Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.679616 4871 scope.go:117] "RemoveContainer" containerID="747581504c4cb242ca1b5b8ac1ab72c6f8680a6a27a5e1f89f29c995e71d07fa" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.926052 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.939362 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.958193 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:29:55 crc kubenswrapper[4871]: E1007 22:29:55.958617 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerName="cinder-api" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.958634 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerName="cinder-api" Oct 07 22:29:55 crc kubenswrapper[4871]: E1007 22:29:55.958646 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.958653 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api" Oct 07 22:29:55 crc kubenswrapper[4871]: E1007 22:29:55.958686 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api-log" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.958692 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api-log" Oct 07 22:29:55 crc kubenswrapper[4871]: E1007 22:29:55.958711 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerName="cinder-api-log" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.958719 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerName="cinder-api-log" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.962938 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerName="cinder-api" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.962970 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" containerName="cinder-api-log" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.962985 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api-log" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.963007 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="998a283e-c92c-4707-af67-a200119a8d75" containerName="barbican-api" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.964061 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.964161 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.966715 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.968214 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.972086 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 07 22:29:55 crc kubenswrapper[4871]: I1007 22:29:55.974432 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.013163 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-scripts\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.013555 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7748t\" (UniqueName: \"kubernetes.io/projected/9d299e74-bffe-4773-a33a-d615606ecc85-kube-api-access-7748t\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.013680 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d299e74-bffe-4773-a33a-d615606ecc85-logs\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.013777 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.013897 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.013971 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.014058 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.014194 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d299e74-bffe-4773-a33a-d615606ecc85-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.014299 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data-custom\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.116122 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-scripts\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.116180 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7748t\" (UniqueName: \"kubernetes.io/projected/9d299e74-bffe-4773-a33a-d615606ecc85-kube-api-access-7748t\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.116224 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d299e74-bffe-4773-a33a-d615606ecc85-logs\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.116253 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.116295 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.116316 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.116343 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.116390 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d299e74-bffe-4773-a33a-d615606ecc85-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.116422 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data-custom\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.117552 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d299e74-bffe-4773-a33a-d615606ecc85-logs\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.117662 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d299e74-bffe-4773-a33a-d615606ecc85-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.122414 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.122999 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.123959 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data-custom\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.124908 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.124920 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.128361 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-scripts\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.135942 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7748t\" (UniqueName: \"kubernetes.io/projected/9d299e74-bffe-4773-a33a-d615606ecc85-kube-api-access-7748t\") pod \"cinder-api-0\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.297586 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 22:29:56 crc kubenswrapper[4871]: I1007 22:29:56.844172 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:29:57 crc kubenswrapper[4871]: I1007 22:29:56.999710 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4" path="/var/lib/kubelet/pods/3a2bb89b-204f-4dfc-bdb5-a995bc75e1d4/volumes" Oct 07 22:29:57 crc kubenswrapper[4871]: I1007 22:29:57.004502 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="998a283e-c92c-4707-af67-a200119a8d75" path="/var/lib/kubelet/pods/998a283e-c92c-4707-af67-a200119a8d75/volumes" Oct 07 22:29:57 crc kubenswrapper[4871]: I1007 22:29:57.599929 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9d299e74-bffe-4773-a33a-d615606ecc85","Type":"ContainerStarted","Data":"8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047"} Oct 07 22:29:57 crc kubenswrapper[4871]: I1007 22:29:57.600263 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9d299e74-bffe-4773-a33a-d615606ecc85","Type":"ContainerStarted","Data":"a4d1e01c9e86249541f40c22a00c33c59fa5c2d70e5a004c774be47711b9bc1a"} Oct 07 22:29:58 crc kubenswrapper[4871]: I1007 22:29:58.617990 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9d299e74-bffe-4773-a33a-d615606ecc85","Type":"ContainerStarted","Data":"b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee"} Oct 07 22:29:58 crc kubenswrapper[4871]: I1007 22:29:58.619477 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 07 22:29:58 crc kubenswrapper[4871]: I1007 22:29:58.648271 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.648217741 podStartE2EDuration="3.648217741s" podCreationTimestamp="2025-10-07 22:29:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:29:58.636402648 +0000 UTC m=+1272.439100721" watchObservedRunningTime="2025-10-07 22:29:58.648217741 +0000 UTC m=+1272.450915814" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.431746 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7748587bfc-lxnz6"] Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.433413 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.438383 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.440038 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.441254 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.449669 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7748587bfc-lxnz6"] Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.505084 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.596321 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-config-data\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.596428 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-etc-swift\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.596471 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-run-httpd\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.596495 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rxdn\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-kube-api-access-7rxdn\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.596540 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-log-httpd\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.596625 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-public-tls-certs\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.596689 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-internal-tls-certs\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.596726 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-combined-ca-bundle\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.698627 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-public-tls-certs\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.698716 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-internal-tls-certs\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.698745 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-combined-ca-bundle\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.698777 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-config-data\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.700270 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-etc-swift\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.700313 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-run-httpd\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.700334 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rxdn\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-kube-api-access-7rxdn\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.700369 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-log-httpd\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.701398 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-run-httpd\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.703057 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-log-httpd\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.712652 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-public-tls-certs\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.713850 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-internal-tls-certs\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.716764 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-config-data\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.719696 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rxdn\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-kube-api-access-7rxdn\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.720394 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-combined-ca-bundle\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.738776 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-etc-swift\") pod \"swift-proxy-7748587bfc-lxnz6\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:29:59 crc kubenswrapper[4871]: I1007 22:29:59.759245 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.152675 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc"] Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.155475 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.160399 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.160840 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.164941 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc"] Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.313702 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21aa285c-6d29-4ef3-b846-bc1f02fc5416-config-volume\") pod \"collect-profiles-29331270-d9xzc\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.315006 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21aa285c-6d29-4ef3-b846-bc1f02fc5416-secret-volume\") pod \"collect-profiles-29331270-d9xzc\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.315044 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpq75\" (UniqueName: \"kubernetes.io/projected/21aa285c-6d29-4ef3-b846-bc1f02fc5416-kube-api-access-cpq75\") pod \"collect-profiles-29331270-d9xzc\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.418066 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21aa285c-6d29-4ef3-b846-bc1f02fc5416-secret-volume\") pod \"collect-profiles-29331270-d9xzc\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.418548 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpq75\" (UniqueName: \"kubernetes.io/projected/21aa285c-6d29-4ef3-b846-bc1f02fc5416-kube-api-access-cpq75\") pod \"collect-profiles-29331270-d9xzc\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.418639 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21aa285c-6d29-4ef3-b846-bc1f02fc5416-config-volume\") pod \"collect-profiles-29331270-d9xzc\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.420081 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21aa285c-6d29-4ef3-b846-bc1f02fc5416-config-volume\") pod \"collect-profiles-29331270-d9xzc\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.429815 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21aa285c-6d29-4ef3-b846-bc1f02fc5416-secret-volume\") pod \"collect-profiles-29331270-d9xzc\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.458340 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpq75\" (UniqueName: \"kubernetes.io/projected/21aa285c-6d29-4ef3-b846-bc1f02fc5416-kube-api-access-cpq75\") pod \"collect-profiles-29331270-d9xzc\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:00 crc kubenswrapper[4871]: I1007 22:30:00.489909 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.026247 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.125241 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-pr26l"] Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.126013 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cddb74997-pr26l" podUID="8fa26949-291d-4703-b9f0-b31858214f37" containerName="dnsmasq-dns" containerID="cri-o://e48e3b09bb6b842b3e6529b901cd5fdd225e0c407aca4b30c7dce97f207717ab" gracePeriod=10 Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.323383 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.400532 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.653146 4871 generic.go:334] "Generic (PLEG): container finished" podID="8fa26949-291d-4703-b9f0-b31858214f37" containerID="e48e3b09bb6b842b3e6529b901cd5fdd225e0c407aca4b30c7dce97f207717ab" exitCode=0 Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.653239 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-pr26l" event={"ID":"8fa26949-291d-4703-b9f0-b31858214f37","Type":"ContainerDied","Data":"e48e3b09bb6b842b3e6529b901cd5fdd225e0c407aca4b30c7dce97f207717ab"} Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.653451 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerName="probe" containerID="cri-o://48ac3b9150034988c91416bc40d56e8a227416f00586e175b4460e93eb981b0f" gracePeriod=30 Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.653565 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerName="cinder-scheduler" containerID="cri-o://c5d24655de275a06d27468ec9a99b8b3dbac30cf3e541319a66a8b4b54758f3e" gracePeriod=30 Oct 07 22:30:01 crc kubenswrapper[4871]: I1007 22:30:01.933538 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:30:02 crc kubenswrapper[4871]: I1007 22:30:02.002230 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-74676f67cb-m997h"] Oct 07 22:30:02 crc kubenswrapper[4871]: I1007 22:30:02.002552 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-74676f67cb-m997h" podUID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerName="neutron-api" containerID="cri-o://8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105" gracePeriod=30 Oct 07 22:30:02 crc kubenswrapper[4871]: I1007 22:30:02.003260 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-74676f67cb-m997h" podUID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerName="neutron-httpd" containerID="cri-o://3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f" gracePeriod=30 Oct 07 22:30:02 crc kubenswrapper[4871]: I1007 22:30:02.669953 4871 generic.go:334] "Generic (PLEG): container finished" podID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerID="3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f" exitCode=0 Oct 07 22:30:02 crc kubenswrapper[4871]: I1007 22:30:02.670026 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74676f67cb-m997h" event={"ID":"c4635daf-677e-4d01-adff-95f9aa80fa7e","Type":"ContainerDied","Data":"3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f"} Oct 07 22:30:02 crc kubenswrapper[4871]: I1007 22:30:02.672262 4871 generic.go:334] "Generic (PLEG): container finished" podID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerID="48ac3b9150034988c91416bc40d56e8a227416f00586e175b4460e93eb981b0f" exitCode=0 Oct 07 22:30:02 crc kubenswrapper[4871]: I1007 22:30:02.672288 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ec31882-e098-41e6-8c3f-f736f8092f9a","Type":"ContainerDied","Data":"48ac3b9150034988c91416bc40d56e8a227416f00586e175b4460e93eb981b0f"} Oct 07 22:30:03 crc kubenswrapper[4871]: I1007 22:30:03.731604 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cddb74997-pr26l" podUID="8fa26949-291d-4703-b9f0-b31858214f37" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.156:5353: connect: connection refused" Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.395230 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.395589 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="ceilometer-central-agent" containerID="cri-o://8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be" gracePeriod=30 Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.395879 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="proxy-httpd" containerID="cri-o://5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40" gracePeriod=30 Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.396181 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="ceilometer-notification-agent" containerID="cri-o://064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334" gracePeriod=30 Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.396206 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="sg-core" containerID="cri-o://6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d" gracePeriod=30 Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.408646 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.159:3000/\": EOF" Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.598383 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.598740 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-log" containerID="cri-o://029c5ef52594f3600fed24649c374ab278a7e185161657daa6156f34dffcf1a9" gracePeriod=30 Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.598764 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-httpd" containerID="cri-o://17ad90817b0347a5b221ac9e0855d33dc50476e984f82363e5299dd2e53ebf52" gracePeriod=30 Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.703351 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerDied","Data":"5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40"} Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.703301 4871 generic.go:334] "Generic (PLEG): container finished" podID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerID="5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40" exitCode=0 Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.703434 4871 generic.go:334] "Generic (PLEG): container finished" podID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerID="6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d" exitCode=2 Oct 07 22:30:04 crc kubenswrapper[4871]: I1007 22:30:04.703458 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerDied","Data":"6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d"} Oct 07 22:30:05 crc kubenswrapper[4871]: I1007 22:30:05.512349 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:30:05 crc kubenswrapper[4871]: I1007 22:30:05.512719 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:30:05 crc kubenswrapper[4871]: I1007 22:30:05.719322 4871 generic.go:334] "Generic (PLEG): container finished" podID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerID="029c5ef52594f3600fed24649c374ab278a7e185161657daa6156f34dffcf1a9" exitCode=143 Oct 07 22:30:05 crc kubenswrapper[4871]: I1007 22:30:05.719370 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb34122e-91f6-4318-8666-f85ea6fd8fbc","Type":"ContainerDied","Data":"029c5ef52594f3600fed24649c374ab278a7e185161657daa6156f34dffcf1a9"} Oct 07 22:30:05 crc kubenswrapper[4871]: I1007 22:30:05.724319 4871 generic.go:334] "Generic (PLEG): container finished" podID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerID="8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be" exitCode=0 Oct 07 22:30:05 crc kubenswrapper[4871]: I1007 22:30:05.724379 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerDied","Data":"8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be"} Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.086989 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.087385 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-log" containerID="cri-o://8dd9a28f4c6c9641ac6b27e2e4d221635516407ec09734b7f97b3fdb49856b47" gracePeriod=30 Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.087475 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-httpd" containerID="cri-o://8173a9d1525d122748f48123b0201f7fa734f3da4bc057a26edb4c400c38a339" gracePeriod=30 Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.097004 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.145:9292/healthcheck\": EOF" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.097205 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/glance-default-internal-api-0" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.145:9292/healthcheck\": EOF" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.665038 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.775860 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-swift-storage-0\") pod \"8fa26949-291d-4703-b9f0-b31858214f37\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.775922 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-nb\") pod \"8fa26949-291d-4703-b9f0-b31858214f37\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.776036 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-config\") pod \"8fa26949-291d-4703-b9f0-b31858214f37\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.776210 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-sb\") pod \"8fa26949-291d-4703-b9f0-b31858214f37\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.776289 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8449c\" (UniqueName: \"kubernetes.io/projected/8fa26949-291d-4703-b9f0-b31858214f37-kube-api-access-8449c\") pod \"8fa26949-291d-4703-b9f0-b31858214f37\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.776381 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-svc\") pod \"8fa26949-291d-4703-b9f0-b31858214f37\" (UID: \"8fa26949-291d-4703-b9f0-b31858214f37\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.778141 4871 generic.go:334] "Generic (PLEG): container finished" podID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerID="c5d24655de275a06d27468ec9a99b8b3dbac30cf3e541319a66a8b4b54758f3e" exitCode=0 Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.778376 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ec31882-e098-41e6-8c3f-f736f8092f9a","Type":"ContainerDied","Data":"c5d24655de275a06d27468ec9a99b8b3dbac30cf3e541319a66a8b4b54758f3e"} Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.810229 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa26949-291d-4703-b9f0-b31858214f37-kube-api-access-8449c" (OuterVolumeSpecName: "kube-api-access-8449c") pod "8fa26949-291d-4703-b9f0-b31858214f37" (UID: "8fa26949-291d-4703-b9f0-b31858214f37"). InnerVolumeSpecName "kube-api-access-8449c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.812243 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-pr26l" event={"ID":"8fa26949-291d-4703-b9f0-b31858214f37","Type":"ContainerDied","Data":"9c354eb3929e4aeb8bf48ab17f0f821f441bd140d110912197afa96c9f9c1d65"} Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.812314 4871 scope.go:117] "RemoveContainer" containerID="e48e3b09bb6b842b3e6529b901cd5fdd225e0c407aca4b30c7dce97f207717ab" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.812544 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cddb74997-pr26l" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.830831 4871 generic.go:334] "Generic (PLEG): container finished" podID="537c4362-75e4-4c06-afff-55499ffed456" containerID="8dd9a28f4c6c9641ac6b27e2e4d221635516407ec09734b7f97b3fdb49856b47" exitCode=143 Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.831014 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"537c4362-75e4-4c06-afff-55499ffed456","Type":"ContainerDied","Data":"8dd9a28f4c6c9641ac6b27e2e4d221635516407ec09734b7f97b3fdb49856b47"} Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.874664 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8fa26949-291d-4703-b9f0-b31858214f37" (UID: "8fa26949-291d-4703-b9f0-b31858214f37"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.881733 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8fa26949-291d-4703-b9f0-b31858214f37" (UID: "8fa26949-291d-4703-b9f0-b31858214f37"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.878527 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8449c\" (UniqueName: \"kubernetes.io/projected/8fa26949-291d-4703-b9f0-b31858214f37-kube-api-access-8449c\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.889860 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-config" (OuterVolumeSpecName: "config") pod "8fa26949-291d-4703-b9f0-b31858214f37" (UID: "8fa26949-291d-4703-b9f0-b31858214f37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.911598 4871 scope.go:117] "RemoveContainer" containerID="0b6868958e3db44b4eb5bfa5d346792a5289cb732b07bd70d215cd1b8ea38450" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.916438 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.920614 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8fa26949-291d-4703-b9f0-b31858214f37" (UID: "8fa26949-291d-4703-b9f0-b31858214f37"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.937711 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8fa26949-291d-4703-b9f0-b31858214f37" (UID: "8fa26949-291d-4703-b9f0-b31858214f37"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.997310 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data-custom\") pod \"5ec31882-e098-41e6-8c3f-f736f8092f9a\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.997417 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data\") pod \"5ec31882-e098-41e6-8c3f-f736f8092f9a\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.997488 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec31882-e098-41e6-8c3f-f736f8092f9a-etc-machine-id\") pod \"5ec31882-e098-41e6-8c3f-f736f8092f9a\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.997530 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-scripts\") pod \"5ec31882-e098-41e6-8c3f-f736f8092f9a\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.997692 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-combined-ca-bundle\") pod \"5ec31882-e098-41e6-8c3f-f736f8092f9a\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.997802 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhdd2\" (UniqueName: \"kubernetes.io/projected/5ec31882-e098-41e6-8c3f-f736f8092f9a-kube-api-access-zhdd2\") pod \"5ec31882-e098-41e6-8c3f-f736f8092f9a\" (UID: \"5ec31882-e098-41e6-8c3f-f736f8092f9a\") " Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.998185 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5ec31882-e098-41e6-8c3f-f736f8092f9a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5ec31882-e098-41e6-8c3f-f736f8092f9a" (UID: "5ec31882-e098-41e6-8c3f-f736f8092f9a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.998682 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.998706 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.998718 4871 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.998730 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.998741 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa26949-291d-4703-b9f0-b31858214f37-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:06 crc kubenswrapper[4871]: I1007 22:30:06.998751 4871 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec31882-e098-41e6-8c3f-f736f8092f9a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.006728 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-scripts" (OuterVolumeSpecName: "scripts") pod "5ec31882-e098-41e6-8c3f-f736f8092f9a" (UID: "5ec31882-e098-41e6-8c3f-f736f8092f9a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.014582 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5ec31882-e098-41e6-8c3f-f736f8092f9a" (UID: "5ec31882-e098-41e6-8c3f-f736f8092f9a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.015438 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ec31882-e098-41e6-8c3f-f736f8092f9a-kube-api-access-zhdd2" (OuterVolumeSpecName: "kube-api-access-zhdd2") pod "5ec31882-e098-41e6-8c3f-f736f8092f9a" (UID: "5ec31882-e098-41e6-8c3f-f736f8092f9a"). InnerVolumeSpecName "kube-api-access-zhdd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.081938 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ec31882-e098-41e6-8c3f-f736f8092f9a" (UID: "5ec31882-e098-41e6-8c3f-f736f8092f9a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.101070 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhdd2\" (UniqueName: \"kubernetes.io/projected/5ec31882-e098-41e6-8c3f-f736f8092f9a-kube-api-access-zhdd2\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.101104 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.101114 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.101124 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.142232 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-pr26l"] Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.159896 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data" (OuterVolumeSpecName: "config-data") pod "5ec31882-e098-41e6-8c3f-f736f8092f9a" (UID: "5ec31882-e098-41e6-8c3f-f736f8092f9a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.162688 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-pr26l"] Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.211174 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec31882-e098-41e6-8c3f-f736f8092f9a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.221209 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc"] Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.559758 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.620578 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-combined-ca-bundle\") pod \"c4635daf-677e-4d01-adff-95f9aa80fa7e\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.621177 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-httpd-config\") pod \"c4635daf-677e-4d01-adff-95f9aa80fa7e\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.621900 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glfqm\" (UniqueName: \"kubernetes.io/projected/c4635daf-677e-4d01-adff-95f9aa80fa7e-kube-api-access-glfqm\") pod \"c4635daf-677e-4d01-adff-95f9aa80fa7e\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.621993 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-ovndb-tls-certs\") pod \"c4635daf-677e-4d01-adff-95f9aa80fa7e\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.622115 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-config\") pod \"c4635daf-677e-4d01-adff-95f9aa80fa7e\" (UID: \"c4635daf-677e-4d01-adff-95f9aa80fa7e\") " Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.630602 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4635daf-677e-4d01-adff-95f9aa80fa7e-kube-api-access-glfqm" (OuterVolumeSpecName: "kube-api-access-glfqm") pod "c4635daf-677e-4d01-adff-95f9aa80fa7e" (UID: "c4635daf-677e-4d01-adff-95f9aa80fa7e"). InnerVolumeSpecName "kube-api-access-glfqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.635027 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c4635daf-677e-4d01-adff-95f9aa80fa7e" (UID: "c4635daf-677e-4d01-adff-95f9aa80fa7e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.691218 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-config" (OuterVolumeSpecName: "config") pod "c4635daf-677e-4d01-adff-95f9aa80fa7e" (UID: "c4635daf-677e-4d01-adff-95f9aa80fa7e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.697223 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4635daf-677e-4d01-adff-95f9aa80fa7e" (UID: "c4635daf-677e-4d01-adff-95f9aa80fa7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.725746 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glfqm\" (UniqueName: \"kubernetes.io/projected/c4635daf-677e-4d01-adff-95f9aa80fa7e-kube-api-access-glfqm\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.725780 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.725804 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.725815 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.734461 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c4635daf-677e-4d01-adff-95f9aa80fa7e" (UID: "c4635daf-677e-4d01-adff-95f9aa80fa7e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.758950 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.146:9292/healthcheck\": read tcp 10.217.0.2:47634->10.217.0.146:9292: read: connection reset by peer" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.759062 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.146:9292/healthcheck\": read tcp 10.217.0.2:47646->10.217.0.146:9292: read: connection reset by peer" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.830997 4871 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4635daf-677e-4d01-adff-95f9aa80fa7e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.842166 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.842146 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ec31882-e098-41e6-8c3f-f736f8092f9a","Type":"ContainerDied","Data":"52751a3dc502aa5cce24e5058c38e2cd54e853850c6485d21e13ab4ff5bec8f7"} Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.842385 4871 scope.go:117] "RemoveContainer" containerID="48ac3b9150034988c91416bc40d56e8a227416f00586e175b4460e93eb981b0f" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.849683 4871 generic.go:334] "Generic (PLEG): container finished" podID="21aa285c-6d29-4ef3-b846-bc1f02fc5416" containerID="d91dafd4d911f590589061e8ce31ae1a7587033551714a110604605d111388b2" exitCode=0 Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.849762 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" event={"ID":"21aa285c-6d29-4ef3-b846-bc1f02fc5416","Type":"ContainerDied","Data":"d91dafd4d911f590589061e8ce31ae1a7587033551714a110604605d111388b2"} Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.849811 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" event={"ID":"21aa285c-6d29-4ef3-b846-bc1f02fc5416","Type":"ContainerStarted","Data":"c6bc49428f6ad2172367c3b1df0791a3e488845cf63111303c735ee12dd9c113"} Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.862625 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"943e025c-54d8-4686-81ee-894259d7af56","Type":"ContainerStarted","Data":"0bf2fc0c8cbcc24b03cb64efbb163db4f05f24420ea760d649a0be66f64959ec"} Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.886893 4871 scope.go:117] "RemoveContainer" containerID="c5d24655de275a06d27468ec9a99b8b3dbac30cf3e541319a66a8b4b54758f3e" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.899526 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.7087561989999998 podStartE2EDuration="14.899503611s" podCreationTimestamp="2025-10-07 22:29:53 +0000 UTC" firstStartedPulling="2025-10-07 22:29:54.239788261 +0000 UTC m=+1268.042486334" lastFinishedPulling="2025-10-07 22:30:06.430535673 +0000 UTC m=+1280.233233746" observedRunningTime="2025-10-07 22:30:07.894350634 +0000 UTC m=+1281.697048707" watchObservedRunningTime="2025-10-07 22:30:07.899503611 +0000 UTC m=+1281.702201684" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.901538 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7748587bfc-lxnz6"] Oct 07 22:30:07 crc kubenswrapper[4871]: W1007 22:30:07.908324 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod645a6027_7304_4fdc_aaf4_ffd433bec275.slice/crio-ac679d9e737a4894873c1cb1029ff1234f15dd5986c1ace6c6d5154d10fd88bf WatchSource:0}: Error finding container ac679d9e737a4894873c1cb1029ff1234f15dd5986c1ace6c6d5154d10fd88bf: Status 404 returned error can't find the container with id ac679d9e737a4894873c1cb1029ff1234f15dd5986c1ace6c6d5154d10fd88bf Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.924632 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.925628 4871 generic.go:334] "Generic (PLEG): container finished" podID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerID="8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105" exitCode=0 Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.925771 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74676f67cb-m997h" event={"ID":"c4635daf-677e-4d01-adff-95f9aa80fa7e","Type":"ContainerDied","Data":"8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105"} Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.925847 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74676f67cb-m997h" event={"ID":"c4635daf-677e-4d01-adff-95f9aa80fa7e","Type":"ContainerDied","Data":"f5f2feb3d9dc5a1c2284ad75b1ff747d1312a2abe5db3864e426981506ad64d9"} Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.926022 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74676f67cb-m997h" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.935934 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.951349 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:30:07 crc kubenswrapper[4871]: E1007 22:30:07.952009 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerName="probe" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952038 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerName="probe" Oct 07 22:30:07 crc kubenswrapper[4871]: E1007 22:30:07.952059 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa26949-291d-4703-b9f0-b31858214f37" containerName="dnsmasq-dns" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952068 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa26949-291d-4703-b9f0-b31858214f37" containerName="dnsmasq-dns" Oct 07 22:30:07 crc kubenswrapper[4871]: E1007 22:30:07.952096 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerName="cinder-scheduler" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952105 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerName="cinder-scheduler" Oct 07 22:30:07 crc kubenswrapper[4871]: E1007 22:30:07.952127 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerName="neutron-api" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952137 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerName="neutron-api" Oct 07 22:30:07 crc kubenswrapper[4871]: E1007 22:30:07.952150 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa26949-291d-4703-b9f0-b31858214f37" containerName="init" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952157 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa26949-291d-4703-b9f0-b31858214f37" containerName="init" Oct 07 22:30:07 crc kubenswrapper[4871]: E1007 22:30:07.952182 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerName="neutron-httpd" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952190 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerName="neutron-httpd" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952422 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerName="cinder-scheduler" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952451 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec31882-e098-41e6-8c3f-f736f8092f9a" containerName="probe" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952466 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa26949-291d-4703-b9f0-b31858214f37" containerName="dnsmasq-dns" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952498 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerName="neutron-httpd" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.952512 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4635daf-677e-4d01-adff-95f9aa80fa7e" containerName="neutron-api" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.953637 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.954153 4871 generic.go:334] "Generic (PLEG): container finished" podID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerID="17ad90817b0347a5b221ac9e0855d33dc50476e984f82363e5299dd2e53ebf52" exitCode=0 Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.954211 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb34122e-91f6-4318-8666-f85ea6fd8fbc","Type":"ContainerDied","Data":"17ad90817b0347a5b221ac9e0855d33dc50476e984f82363e5299dd2e53ebf52"} Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.956941 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.971317 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.981054 4871 scope.go:117] "RemoveContainer" containerID="3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f" Oct 07 22:30:07 crc kubenswrapper[4871]: I1007 22:30:07.997894 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-74676f67cb-m997h"] Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.014688 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-74676f67cb-m997h"] Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.041574 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-scripts\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.041639 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.041667 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.041685 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.041745 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/825fba74-f028-4445-b041-465ccbbad763-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.041806 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg9gt\" (UniqueName: \"kubernetes.io/projected/825fba74-f028-4445-b041-465ccbbad763-kube-api-access-jg9gt\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.069980 4871 scope.go:117] "RemoveContainer" containerID="8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.137276 4871 scope.go:117] "RemoveContainer" containerID="3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f" Oct 07 22:30:08 crc kubenswrapper[4871]: E1007 22:30:08.138287 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f\": container with ID starting with 3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f not found: ID does not exist" containerID="3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.138338 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f"} err="failed to get container status \"3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f\": rpc error: code = NotFound desc = could not find container \"3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f\": container with ID starting with 3ac4218d3b6685f7f5cc49b4bae87cec68d276b1267b7195ad8feae8fa0afa4f not found: ID does not exist" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.138364 4871 scope.go:117] "RemoveContainer" containerID="8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.145577 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg9gt\" (UniqueName: \"kubernetes.io/projected/825fba74-f028-4445-b041-465ccbbad763-kube-api-access-jg9gt\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.145683 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-scripts\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.145718 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.145736 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.145752 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.145830 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/825fba74-f028-4445-b041-465ccbbad763-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.145928 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/825fba74-f028-4445-b041-465ccbbad763-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: E1007 22:30:08.145992 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105\": container with ID starting with 8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105 not found: ID does not exist" containerID="8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.146018 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105"} err="failed to get container status \"8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105\": rpc error: code = NotFound desc = could not find container \"8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105\": container with ID starting with 8a15a30a40c93e71a4fb3f3e6be39dc8336e327d0e4a3405538e184936b27105 not found: ID does not exist" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.158223 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-scripts\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.158494 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.158750 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.167187 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg9gt\" (UniqueName: \"kubernetes.io/projected/825fba74-f028-4445-b041-465ccbbad763-kube-api-access-jg9gt\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.171417 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data\") pod \"cinder-scheduler-0\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.351469 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.542353 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.664546 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvcwv\" (UniqueName: \"kubernetes.io/projected/fb34122e-91f6-4318-8666-f85ea6fd8fbc-kube-api-access-dvcwv\") pod \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.665025 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-httpd-run\") pod \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.665143 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.665176 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-public-tls-certs\") pod \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.665218 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-config-data\") pod \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.665264 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-logs\") pod \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.665293 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-scripts\") pod \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.665346 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-combined-ca-bundle\") pod \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\" (UID: \"fb34122e-91f6-4318-8666-f85ea6fd8fbc\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.665634 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fb34122e-91f6-4318-8666-f85ea6fd8fbc" (UID: "fb34122e-91f6-4318-8666-f85ea6fd8fbc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.666335 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.668659 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-logs" (OuterVolumeSpecName: "logs") pod "fb34122e-91f6-4318-8666-f85ea6fd8fbc" (UID: "fb34122e-91f6-4318-8666-f85ea6fd8fbc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.678696 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb34122e-91f6-4318-8666-f85ea6fd8fbc-kube-api-access-dvcwv" (OuterVolumeSpecName: "kube-api-access-dvcwv") pod "fb34122e-91f6-4318-8666-f85ea6fd8fbc" (UID: "fb34122e-91f6-4318-8666-f85ea6fd8fbc"). InnerVolumeSpecName "kube-api-access-dvcwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.689817 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-scripts" (OuterVolumeSpecName: "scripts") pod "fb34122e-91f6-4318-8666-f85ea6fd8fbc" (UID: "fb34122e-91f6-4318-8666-f85ea6fd8fbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.690063 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "fb34122e-91f6-4318-8666-f85ea6fd8fbc" (UID: "fb34122e-91f6-4318-8666-f85ea6fd8fbc"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.717162 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb34122e-91f6-4318-8666-f85ea6fd8fbc" (UID: "fb34122e-91f6-4318-8666-f85ea6fd8fbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.769771 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb34122e-91f6-4318-8666-f85ea6fd8fbc-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.770066 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.770078 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.770097 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvcwv\" (UniqueName: \"kubernetes.io/projected/fb34122e-91f6-4318-8666-f85ea6fd8fbc-kube-api-access-dvcwv\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.770130 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.773459 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.820189 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fb34122e-91f6-4318-8666-f85ea6fd8fbc" (UID: "fb34122e-91f6-4318-8666-f85ea6fd8fbc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.830993 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-config-data" (OuterVolumeSpecName: "config-data") pod "fb34122e-91f6-4318-8666-f85ea6fd8fbc" (UID: "fb34122e-91f6-4318-8666-f85ea6fd8fbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.839022 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.871808 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-combined-ca-bundle\") pod \"ae96a676-3a8b-44f0-a54a-5031450388d1\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.871916 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v9cc\" (UniqueName: \"kubernetes.io/projected/ae96a676-3a8b-44f0-a54a-5031450388d1-kube-api-access-8v9cc\") pod \"ae96a676-3a8b-44f0-a54a-5031450388d1\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.872060 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-log-httpd\") pod \"ae96a676-3a8b-44f0-a54a-5031450388d1\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.872084 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-config-data\") pod \"ae96a676-3a8b-44f0-a54a-5031450388d1\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.872321 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-run-httpd\") pod \"ae96a676-3a8b-44f0-a54a-5031450388d1\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.872341 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-scripts\") pod \"ae96a676-3a8b-44f0-a54a-5031450388d1\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.872363 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-sg-core-conf-yaml\") pod \"ae96a676-3a8b-44f0-a54a-5031450388d1\" (UID: \"ae96a676-3a8b-44f0-a54a-5031450388d1\") " Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.873193 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.873259 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.873276 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb34122e-91f6-4318-8666-f85ea6fd8fbc-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.874619 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ae96a676-3a8b-44f0-a54a-5031450388d1" (UID: "ae96a676-3a8b-44f0-a54a-5031450388d1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.874663 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ae96a676-3a8b-44f0-a54a-5031450388d1" (UID: "ae96a676-3a8b-44f0-a54a-5031450388d1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.878679 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae96a676-3a8b-44f0-a54a-5031450388d1-kube-api-access-8v9cc" (OuterVolumeSpecName: "kube-api-access-8v9cc") pod "ae96a676-3a8b-44f0-a54a-5031450388d1" (UID: "ae96a676-3a8b-44f0-a54a-5031450388d1"). InnerVolumeSpecName "kube-api-access-8v9cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.878725 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-scripts" (OuterVolumeSpecName: "scripts") pod "ae96a676-3a8b-44f0-a54a-5031450388d1" (UID: "ae96a676-3a8b-44f0-a54a-5031450388d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.918735 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ae96a676-3a8b-44f0-a54a-5031450388d1" (UID: "ae96a676-3a8b-44f0-a54a-5031450388d1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.975774 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.975820 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.975830 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.975844 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v9cc\" (UniqueName: \"kubernetes.io/projected/ae96a676-3a8b-44f0-a54a-5031450388d1-kube-api-access-8v9cc\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.975853 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae96a676-3a8b-44f0-a54a-5031450388d1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.976485 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7748587bfc-lxnz6" event={"ID":"645a6027-7304-4fdc-aaf4-ffd433bec275","Type":"ContainerStarted","Data":"24dbb24a65b2912e92cb5795b73a40b47b50468929436754aa78a71953f92649"} Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.976535 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7748587bfc-lxnz6" event={"ID":"645a6027-7304-4fdc-aaf4-ffd433bec275","Type":"ContainerStarted","Data":"8348036e75afc1a3c2a675f95267987dea996bddef522f920d21ab3ccc437dfd"} Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.976551 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7748587bfc-lxnz6" event={"ID":"645a6027-7304-4fdc-aaf4-ffd433bec275","Type":"ContainerStarted","Data":"ac679d9e737a4894873c1cb1029ff1234f15dd5986c1ace6c6d5154d10fd88bf"} Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.977410 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.977504 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.981742 4871 generic.go:334] "Generic (PLEG): container finished" podID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerID="064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334" exitCode=0 Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.981907 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.986883 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae96a676-3a8b-44f0-a54a-5031450388d1" (UID: "ae96a676-3a8b-44f0-a54a-5031450388d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:08 crc kubenswrapper[4871]: I1007 22:30:08.990473 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.002243 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-config-data" (OuterVolumeSpecName: "config-data") pod "ae96a676-3a8b-44f0-a54a-5031450388d1" (UID: "ae96a676-3a8b-44f0-a54a-5031450388d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.008967 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7748587bfc-lxnz6" podStartSLOduration=10.008912954 podStartE2EDuration="10.008912954s" podCreationTimestamp="2025-10-07 22:29:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:30:09.002170135 +0000 UTC m=+1282.804868208" watchObservedRunningTime="2025-10-07 22:30:09.008912954 +0000 UTC m=+1282.811611027" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.010831 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ec31882-e098-41e6-8c3f-f736f8092f9a" path="/var/lib/kubelet/pods/5ec31882-e098-41e6-8c3f-f736f8092f9a/volumes" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.012241 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fa26949-291d-4703-b9f0-b31858214f37" path="/var/lib/kubelet/pods/8fa26949-291d-4703-b9f0-b31858214f37/volumes" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.013042 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4635daf-677e-4d01-adff-95f9aa80fa7e" path="/var/lib/kubelet/pods/c4635daf-677e-4d01-adff-95f9aa80fa7e/volumes" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.018288 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerDied","Data":"064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334"} Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.018345 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae96a676-3a8b-44f0-a54a-5031450388d1","Type":"ContainerDied","Data":"81b8621a149a9b4fdc1c9d267bd6d1227c1fbf52f494d2b3c9fb498c983ab7fd"} Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.018375 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb34122e-91f6-4318-8666-f85ea6fd8fbc","Type":"ContainerDied","Data":"d433d98e82a8966ce766449f7bdea2d5cc7e768065f9bb55a27556194b51bd6b"} Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.018409 4871 scope.go:117] "RemoveContainer" containerID="5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.044545 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.084890 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.084919 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae96a676-3a8b-44f0-a54a-5031450388d1-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.101895 4871 scope.go:117] "RemoveContainer" containerID="6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.106897 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.135959 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.178527 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.194532 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.195009 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="ceilometer-notification-agent" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195030 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="ceilometer-notification-agent" Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.195038 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="proxy-httpd" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195045 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="proxy-httpd" Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.195064 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-log" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195072 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-log" Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.195084 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-httpd" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195090 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-httpd" Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.195106 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="ceilometer-central-agent" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195112 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="ceilometer-central-agent" Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.195124 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="sg-core" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195130 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="sg-core" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195296 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="sg-core" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195312 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="ceilometer-central-agent" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195318 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="ceilometer-notification-agent" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195329 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-httpd" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195337 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" containerName="proxy-httpd" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.195356 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" containerName="glance-log" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.196840 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.201470 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.202672 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.203642 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.219289 4871 scope.go:117] "RemoveContainer" containerID="064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.352173 4871 scope.go:117] "RemoveContainer" containerID="8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.390495 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.405484 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.406381 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4rcl\" (UniqueName: \"kubernetes.io/projected/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-kube-api-access-r4rcl\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.406413 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-scripts\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.406483 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-logs\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.406506 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.406525 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.406551 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-config-data\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.406583 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.406637 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.445964 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.448519 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.458814 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.473272 4871 scope.go:117] "RemoveContainer" containerID="5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.473427 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.473480 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.475930 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40\": container with ID starting with 5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40 not found: ID does not exist" containerID="5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.476080 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40"} err="failed to get container status \"5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40\": rpc error: code = NotFound desc = could not find container \"5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40\": container with ID starting with 5ad58880c0e05799942c36dc877da959b610a7a55682365458cb5f80ad32ae40 not found: ID does not exist" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.476224 4871 scope.go:117] "RemoveContainer" containerID="6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d" Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.486592 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d\": container with ID starting with 6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d not found: ID does not exist" containerID="6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.486640 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d"} err="failed to get container status \"6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d\": rpc error: code = NotFound desc = could not find container \"6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d\": container with ID starting with 6b539034fd948cf89689fbd3e899e7ccb78adde1f31a53adf4c038e664360b5d not found: ID does not exist" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.486675 4871 scope.go:117] "RemoveContainer" containerID="064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334" Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.498220 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334\": container with ID starting with 064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334 not found: ID does not exist" containerID="064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.498560 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334"} err="failed to get container status \"064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334\": rpc error: code = NotFound desc = could not find container \"064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334\": container with ID starting with 064065b5fc288575a04993738d249f1f70901a0e3dd2dabfede6d8b79fc10334 not found: ID does not exist" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.498683 4871 scope.go:117] "RemoveContainer" containerID="8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be" Oct 07 22:30:09 crc kubenswrapper[4871]: E1007 22:30:09.501954 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be\": container with ID starting with 8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be not found: ID does not exist" containerID="8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.502012 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be"} err="failed to get container status \"8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be\": rpc error: code = NotFound desc = could not find container \"8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be\": container with ID starting with 8eae168d7cfebb18dd2bd1b85eee49eda403e84058ef344848ff5da0d25903be not found: ID does not exist" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.502048 4871 scope.go:117] "RemoveContainer" containerID="17ad90817b0347a5b221ac9e0855d33dc50476e984f82363e5299dd2e53ebf52" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.509915 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-logs\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.510339 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.510412 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.510511 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-config-data\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.510603 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.510734 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.510808 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-logs\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.510887 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4rcl\" (UniqueName: \"kubernetes.io/projected/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-kube-api-access-r4rcl\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.510978 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-scripts\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.511457 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.514543 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.522457 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-scripts\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.534514 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.535492 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-config-data\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.536109 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.538712 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4rcl\" (UniqueName: \"kubernetes.io/projected/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-kube-api-access-r4rcl\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.556171 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.597458 4871 scope.go:117] "RemoveContainer" containerID="029c5ef52594f3600fed24649c374ab278a7e185161657daa6156f34dffcf1a9" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.617308 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-config-data\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.617381 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-scripts\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.618086 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-run-httpd\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.618157 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.618187 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-log-httpd\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.618281 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh4x2\" (UniqueName: \"kubernetes.io/projected/551d28e0-eb30-4837-9130-d53f9ce982ac-kube-api-access-jh4x2\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.618322 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.630519 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.706618 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.720298 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21aa285c-6d29-4ef3-b846-bc1f02fc5416-config-volume\") pod \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.720408 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpq75\" (UniqueName: \"kubernetes.io/projected/21aa285c-6d29-4ef3-b846-bc1f02fc5416-kube-api-access-cpq75\") pod \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.720439 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21aa285c-6d29-4ef3-b846-bc1f02fc5416-secret-volume\") pod \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\" (UID: \"21aa285c-6d29-4ef3-b846-bc1f02fc5416\") " Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.720720 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh4x2\" (UniqueName: \"kubernetes.io/projected/551d28e0-eb30-4837-9130-d53f9ce982ac-kube-api-access-jh4x2\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.720761 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.720834 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-config-data\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.720897 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-scripts\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.720920 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-run-httpd\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.720978 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.721003 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-log-httpd\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.721441 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21aa285c-6d29-4ef3-b846-bc1f02fc5416-config-volume" (OuterVolumeSpecName: "config-volume") pod "21aa285c-6d29-4ef3-b846-bc1f02fc5416" (UID: "21aa285c-6d29-4ef3-b846-bc1f02fc5416"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.721563 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-log-httpd\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.723597 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-run-httpd\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.725457 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21aa285c-6d29-4ef3-b846-bc1f02fc5416-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "21aa285c-6d29-4ef3-b846-bc1f02fc5416" (UID: "21aa285c-6d29-4ef3-b846-bc1f02fc5416"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.728214 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-config-data\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.728811 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-scripts\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.728820 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.731951 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21aa285c-6d29-4ef3-b846-bc1f02fc5416-kube-api-access-cpq75" (OuterVolumeSpecName: "kube-api-access-cpq75") pod "21aa285c-6d29-4ef3-b846-bc1f02fc5416" (UID: "21aa285c-6d29-4ef3-b846-bc1f02fc5416"). InnerVolumeSpecName "kube-api-access-cpq75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.732208 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.738291 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh4x2\" (UniqueName: \"kubernetes.io/projected/551d28e0-eb30-4837-9130-d53f9ce982ac-kube-api-access-jh4x2\") pod \"ceilometer-0\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.815123 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.822704 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21aa285c-6d29-4ef3-b846-bc1f02fc5416-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.822744 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpq75\" (UniqueName: \"kubernetes.io/projected/21aa285c-6d29-4ef3-b846-bc1f02fc5416-kube-api-access-cpq75\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:09 crc kubenswrapper[4871]: I1007 22:30:09.822757 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21aa285c-6d29-4ef3-b846-bc1f02fc5416-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:10 crc kubenswrapper[4871]: I1007 22:30:10.024940 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" Oct 07 22:30:10 crc kubenswrapper[4871]: I1007 22:30:10.025100 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc" event={"ID":"21aa285c-6d29-4ef3-b846-bc1f02fc5416","Type":"ContainerDied","Data":"c6bc49428f6ad2172367c3b1df0791a3e488845cf63111303c735ee12dd9c113"} Oct 07 22:30:10 crc kubenswrapper[4871]: I1007 22:30:10.025980 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6bc49428f6ad2172367c3b1df0791a3e488845cf63111303c735ee12dd9c113" Oct 07 22:30:10 crc kubenswrapper[4871]: I1007 22:30:10.039643 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"825fba74-f028-4445-b041-465ccbbad763","Type":"ContainerStarted","Data":"00a248eb0776a91ccd453b5b36dc1d36cec9eadd09664079556cc912f23bbbd8"} Oct 07 22:30:10 crc kubenswrapper[4871]: I1007 22:30:10.409881 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:30:10 crc kubenswrapper[4871]: I1007 22:30:10.421747 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.002558 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae96a676-3a8b-44f0-a54a-5031450388d1" path="/var/lib/kubelet/pods/ae96a676-3a8b-44f0-a54a-5031450388d1/volumes" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.005585 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb34122e-91f6-4318-8666-f85ea6fd8fbc" path="/var/lib/kubelet/pods/fb34122e-91f6-4318-8666-f85ea6fd8fbc/volumes" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.064280 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"825fba74-f028-4445-b041-465ccbbad763","Type":"ContainerStarted","Data":"ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9"} Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.072444 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f1ac984d-67d5-4bd1-b3c2-ef99976a8708","Type":"ContainerStarted","Data":"b1edbe024bdd7b1061a46862e4599c8dafd94febc9e77e27c28eb7de224466f9"} Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.077146 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerStarted","Data":"e646a44b91fbd3cb54ae458e04cd877aee83c0b7abdc2c57a7003da4257054cb"} Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.086250 4871 generic.go:334] "Generic (PLEG): container finished" podID="537c4362-75e4-4c06-afff-55499ffed456" containerID="8173a9d1525d122748f48123b0201f7fa734f3da4bc057a26edb4c400c38a339" exitCode=0 Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.086646 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"537c4362-75e4-4c06-afff-55499ffed456","Type":"ContainerDied","Data":"8173a9d1525d122748f48123b0201f7fa734f3da4bc057a26edb4c400c38a339"} Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.354763 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.491525 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-httpd-run\") pod \"537c4362-75e4-4c06-afff-55499ffed456\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.492069 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-logs\") pod \"537c4362-75e4-4c06-afff-55499ffed456\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.492113 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"537c4362-75e4-4c06-afff-55499ffed456\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.492179 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-scripts\") pod \"537c4362-75e4-4c06-afff-55499ffed456\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.492308 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxms8\" (UniqueName: \"kubernetes.io/projected/537c4362-75e4-4c06-afff-55499ffed456-kube-api-access-vxms8\") pod \"537c4362-75e4-4c06-afff-55499ffed456\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.492380 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-combined-ca-bundle\") pod \"537c4362-75e4-4c06-afff-55499ffed456\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.492436 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-config-data\") pod \"537c4362-75e4-4c06-afff-55499ffed456\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.492520 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-internal-tls-certs\") pod \"537c4362-75e4-4c06-afff-55499ffed456\" (UID: \"537c4362-75e4-4c06-afff-55499ffed456\") " Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.492767 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "537c4362-75e4-4c06-afff-55499ffed456" (UID: "537c4362-75e4-4c06-afff-55499ffed456"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.493469 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.493855 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-logs" (OuterVolumeSpecName: "logs") pod "537c4362-75e4-4c06-afff-55499ffed456" (UID: "537c4362-75e4-4c06-afff-55499ffed456"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.498348 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/537c4362-75e4-4c06-afff-55499ffed456-kube-api-access-vxms8" (OuterVolumeSpecName: "kube-api-access-vxms8") pod "537c4362-75e4-4c06-afff-55499ffed456" (UID: "537c4362-75e4-4c06-afff-55499ffed456"). InnerVolumeSpecName "kube-api-access-vxms8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.505988 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-scripts" (OuterVolumeSpecName: "scripts") pod "537c4362-75e4-4c06-afff-55499ffed456" (UID: "537c4362-75e4-4c06-afff-55499ffed456"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.520539 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "537c4362-75e4-4c06-afff-55499ffed456" (UID: "537c4362-75e4-4c06-afff-55499ffed456"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.539158 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "537c4362-75e4-4c06-afff-55499ffed456" (UID: "537c4362-75e4-4c06-afff-55499ffed456"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.549864 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-config-data" (OuterVolumeSpecName: "config-data") pod "537c4362-75e4-4c06-afff-55499ffed456" (UID: "537c4362-75e4-4c06-afff-55499ffed456"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.568471 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "537c4362-75e4-4c06-afff-55499ffed456" (UID: "537c4362-75e4-4c06-afff-55499ffed456"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.595983 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxms8\" (UniqueName: \"kubernetes.io/projected/537c4362-75e4-4c06-afff-55499ffed456-kube-api-access-vxms8\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.596022 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.596031 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.596040 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.596049 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/537c4362-75e4-4c06-afff-55499ffed456-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.596096 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.596105 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/537c4362-75e4-4c06-afff-55499ffed456-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.618982 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 07 22:30:11 crc kubenswrapper[4871]: I1007 22:30:11.698704 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.108011 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"537c4362-75e4-4c06-afff-55499ffed456","Type":"ContainerDied","Data":"601b59ada2c4a8f358447ac89f74f115ebceff7fa3a0899001ad2ccb3838426f"} Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.108404 4871 scope.go:117] "RemoveContainer" containerID="8173a9d1525d122748f48123b0201f7fa734f3da4bc057a26edb4c400c38a339" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.108587 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.118113 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"825fba74-f028-4445-b041-465ccbbad763","Type":"ContainerStarted","Data":"b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1"} Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.131769 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f1ac984d-67d5-4bd1-b3c2-ef99976a8708","Type":"ContainerStarted","Data":"a6bcbbe481eb19cf659465c0d4fcb0fa01acc7646b80f2eaea5e2ac2187ebf78"} Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.132594 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f1ac984d-67d5-4bd1-b3c2-ef99976a8708","Type":"ContainerStarted","Data":"0b297eaadbd30609ad30865b36ac820cfdedfdd27b2cf2f2a38dea46f5c786d9"} Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.143961 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerStarted","Data":"0f840f713a6e4e09878020dc4aa7e2d0a2fa4bcf846d4e1c059b75acc895d51a"} Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.180229 4871 scope.go:117] "RemoveContainer" containerID="8dd9a28f4c6c9641ac6b27e2e4d221635516407ec09734b7f97b3fdb49856b47" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.200960 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.200931173 podStartE2EDuration="5.200931173s" podCreationTimestamp="2025-10-07 22:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:30:12.145396411 +0000 UTC m=+1285.948094484" watchObservedRunningTime="2025-10-07 22:30:12.200931173 +0000 UTC m=+1286.003629246" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.202910 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.202891725 podStartE2EDuration="3.202891725s" podCreationTimestamp="2025-10-07 22:30:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:30:12.180900732 +0000 UTC m=+1285.983598805" watchObservedRunningTime="2025-10-07 22:30:12.202891725 +0000 UTC m=+1286.005589788" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.222748 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.244938 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.261683 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:30:12 crc kubenswrapper[4871]: E1007 22:30:12.262669 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-httpd" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.262771 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-httpd" Oct 07 22:30:12 crc kubenswrapper[4871]: E1007 22:30:12.262924 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-log" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.262994 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-log" Oct 07 22:30:12 crc kubenswrapper[4871]: E1007 22:30:12.263107 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21aa285c-6d29-4ef3-b846-bc1f02fc5416" containerName="collect-profiles" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.263175 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="21aa285c-6d29-4ef3-b846-bc1f02fc5416" containerName="collect-profiles" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.263483 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-httpd" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.263578 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="21aa285c-6d29-4ef3-b846-bc1f02fc5416" containerName="collect-profiles" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.263679 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="537c4362-75e4-4c06-afff-55499ffed456" containerName="glance-log" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.265428 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.271242 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.271444 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.291739 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.419656 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.419712 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.419758 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.419808 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.419855 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.420284 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9djzg\" (UniqueName: \"kubernetes.io/projected/d02631b4-f3a6-48d5-a43e-c90a48622ec2-kube-api-access-9djzg\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.420324 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-logs\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.420391 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.522103 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.522446 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.522777 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.522930 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.523027 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.523027 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.523371 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.523394 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9djzg\" (UniqueName: \"kubernetes.io/projected/d02631b4-f3a6-48d5-a43e-c90a48622ec2-kube-api-access-9djzg\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.523431 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-logs\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.523528 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.525873 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-logs\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.532707 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.534893 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.539466 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.543613 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.553336 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9djzg\" (UniqueName: \"kubernetes.io/projected/d02631b4-f3a6-48d5-a43e-c90a48622ec2-kube-api-access-9djzg\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.561170 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.588217 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:12 crc kubenswrapper[4871]: I1007 22:30:12.998677 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="537c4362-75e4-4c06-afff-55499ffed456" path="/var/lib/kubelet/pods/537c4362-75e4-4c06-afff-55499ffed456/volumes" Oct 07 22:30:13 crc kubenswrapper[4871]: I1007 22:30:13.207476 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:30:13 crc kubenswrapper[4871]: W1007 22:30:13.215197 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd02631b4_f3a6_48d5_a43e_c90a48622ec2.slice/crio-5f7c928bf87134511100f2455772d75ebb4dade15c1424ad2a868ec3ac15b223 WatchSource:0}: Error finding container 5f7c928bf87134511100f2455772d75ebb4dade15c1424ad2a868ec3ac15b223: Status 404 returned error can't find the container with id 5f7c928bf87134511100f2455772d75ebb4dade15c1424ad2a868ec3ac15b223 Oct 07 22:30:13 crc kubenswrapper[4871]: I1007 22:30:13.353080 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 07 22:30:13 crc kubenswrapper[4871]: I1007 22:30:13.931722 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:14 crc kubenswrapper[4871]: I1007 22:30:14.178814 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerStarted","Data":"01772163177e6e5575aedbf766fcb515057393c197eccbafc84345eb36f25284"} Oct 07 22:30:14 crc kubenswrapper[4871]: I1007 22:30:14.186420 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d02631b4-f3a6-48d5-a43e-c90a48622ec2","Type":"ContainerStarted","Data":"303e52fe41cb9eb2c55a65b8e6fb2c26f6e89c632f0d230eb1f5ba71eb482fee"} Oct 07 22:30:14 crc kubenswrapper[4871]: I1007 22:30:14.186473 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d02631b4-f3a6-48d5-a43e-c90a48622ec2","Type":"ContainerStarted","Data":"5f7c928bf87134511100f2455772d75ebb4dade15c1424ad2a868ec3ac15b223"} Oct 07 22:30:14 crc kubenswrapper[4871]: I1007 22:30:14.773533 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:30:14 crc kubenswrapper[4871]: I1007 22:30:14.775069 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:30:15 crc kubenswrapper[4871]: I1007 22:30:15.197204 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d02631b4-f3a6-48d5-a43e-c90a48622ec2","Type":"ContainerStarted","Data":"91eb51510e831d42f41a70585ae844bb55e88150c93eae1571a50da925c20bf3"} Oct 07 22:30:15 crc kubenswrapper[4871]: I1007 22:30:15.220053 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.220030899 podStartE2EDuration="3.220030899s" podCreationTimestamp="2025-10-07 22:30:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:30:15.217625276 +0000 UTC m=+1289.020323359" watchObservedRunningTime="2025-10-07 22:30:15.220030899 +0000 UTC m=+1289.022728972" Oct 07 22:30:16 crc kubenswrapper[4871]: I1007 22:30:16.211823 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerStarted","Data":"c3039b7e24a5a3966ed72efc67766c99d2c5cb25a339cf4a7714e6f8c636e1d1"} Oct 07 22:30:18 crc kubenswrapper[4871]: I1007 22:30:18.605211 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.247259 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerStarted","Data":"5a2b0d3c5b9d10377926ce926087215d4e36762a8c11372d05fc82ca33153268"} Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.247496 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="ceilometer-central-agent" containerID="cri-o://0f840f713a6e4e09878020dc4aa7e2d0a2fa4bcf846d4e1c059b75acc895d51a" gracePeriod=30 Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.247582 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="sg-core" containerID="cri-o://c3039b7e24a5a3966ed72efc67766c99d2c5cb25a339cf4a7714e6f8c636e1d1" gracePeriod=30 Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.247665 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="ceilometer-notification-agent" containerID="cri-o://01772163177e6e5575aedbf766fcb515057393c197eccbafc84345eb36f25284" gracePeriod=30 Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.247826 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.247760 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="proxy-httpd" containerID="cri-o://5a2b0d3c5b9d10377926ce926087215d4e36762a8c11372d05fc82ca33153268" gracePeriod=30 Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.285955 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.637031762 podStartE2EDuration="10.285912207s" podCreationTimestamp="2025-10-07 22:30:09 +0000 UTC" firstStartedPulling="2025-10-07 22:30:10.446861478 +0000 UTC m=+1284.249559551" lastFinishedPulling="2025-10-07 22:30:18.095741913 +0000 UTC m=+1291.898439996" observedRunningTime="2025-10-07 22:30:19.270885338 +0000 UTC m=+1293.073583411" watchObservedRunningTime="2025-10-07 22:30:19.285912207 +0000 UTC m=+1293.088610290" Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.707453 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.707612 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.754660 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 22:30:19 crc kubenswrapper[4871]: I1007 22:30:19.807251 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.261136 4871 generic.go:334] "Generic (PLEG): container finished" podID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerID="5a2b0d3c5b9d10377926ce926087215d4e36762a8c11372d05fc82ca33153268" exitCode=0 Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.261591 4871 generic.go:334] "Generic (PLEG): container finished" podID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerID="c3039b7e24a5a3966ed72efc67766c99d2c5cb25a339cf4a7714e6f8c636e1d1" exitCode=2 Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.261603 4871 generic.go:334] "Generic (PLEG): container finished" podID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerID="01772163177e6e5575aedbf766fcb515057393c197eccbafc84345eb36f25284" exitCode=0 Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.261611 4871 generic.go:334] "Generic (PLEG): container finished" podID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerID="0f840f713a6e4e09878020dc4aa7e2d0a2fa4bcf846d4e1c059b75acc895d51a" exitCode=0 Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.261204 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerDied","Data":"5a2b0d3c5b9d10377926ce926087215d4e36762a8c11372d05fc82ca33153268"} Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.261704 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerDied","Data":"c3039b7e24a5a3966ed72efc67766c99d2c5cb25a339cf4a7714e6f8c636e1d1"} Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.261717 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerDied","Data":"01772163177e6e5575aedbf766fcb515057393c197eccbafc84345eb36f25284"} Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.261728 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerDied","Data":"0f840f713a6e4e09878020dc4aa7e2d0a2fa4bcf846d4e1c059b75acc895d51a"} Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.262079 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.262209 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.688099 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.829204 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-scripts\") pod \"551d28e0-eb30-4837-9130-d53f9ce982ac\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.829313 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-log-httpd\") pod \"551d28e0-eb30-4837-9130-d53f9ce982ac\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.829340 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-run-httpd\") pod \"551d28e0-eb30-4837-9130-d53f9ce982ac\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.829389 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-combined-ca-bundle\") pod \"551d28e0-eb30-4837-9130-d53f9ce982ac\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.829417 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-sg-core-conf-yaml\") pod \"551d28e0-eb30-4837-9130-d53f9ce982ac\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.829984 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "551d28e0-eb30-4837-9130-d53f9ce982ac" (UID: "551d28e0-eb30-4837-9130-d53f9ce982ac"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.830148 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "551d28e0-eb30-4837-9130-d53f9ce982ac" (UID: "551d28e0-eb30-4837-9130-d53f9ce982ac"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.830387 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-config-data\") pod \"551d28e0-eb30-4837-9130-d53f9ce982ac\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.830413 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh4x2\" (UniqueName: \"kubernetes.io/projected/551d28e0-eb30-4837-9130-d53f9ce982ac-kube-api-access-jh4x2\") pod \"551d28e0-eb30-4837-9130-d53f9ce982ac\" (UID: \"551d28e0-eb30-4837-9130-d53f9ce982ac\") " Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.830910 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.830925 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/551d28e0-eb30-4837-9130-d53f9ce982ac-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.840314 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-scripts" (OuterVolumeSpecName: "scripts") pod "551d28e0-eb30-4837-9130-d53f9ce982ac" (UID: "551d28e0-eb30-4837-9130-d53f9ce982ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.849758 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/551d28e0-eb30-4837-9130-d53f9ce982ac-kube-api-access-jh4x2" (OuterVolumeSpecName: "kube-api-access-jh4x2") pod "551d28e0-eb30-4837-9130-d53f9ce982ac" (UID: "551d28e0-eb30-4837-9130-d53f9ce982ac"). InnerVolumeSpecName "kube-api-access-jh4x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.865876 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "551d28e0-eb30-4837-9130-d53f9ce982ac" (UID: "551d28e0-eb30-4837-9130-d53f9ce982ac"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.932822 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh4x2\" (UniqueName: \"kubernetes.io/projected/551d28e0-eb30-4837-9130-d53f9ce982ac-kube-api-access-jh4x2\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.932855 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.932865 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.943621 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "551d28e0-eb30-4837-9130-d53f9ce982ac" (UID: "551d28e0-eb30-4837-9130-d53f9ce982ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:20 crc kubenswrapper[4871]: I1007 22:30:20.979684 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-config-data" (OuterVolumeSpecName: "config-data") pod "551d28e0-eb30-4837-9130-d53f9ce982ac" (UID: "551d28e0-eb30-4837-9130-d53f9ce982ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.035117 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.035161 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551d28e0-eb30-4837-9130-d53f9ce982ac-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.278967 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"551d28e0-eb30-4837-9130-d53f9ce982ac","Type":"ContainerDied","Data":"e646a44b91fbd3cb54ae458e04cd877aee83c0b7abdc2c57a7003da4257054cb"} Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.279069 4871 scope.go:117] "RemoveContainer" containerID="5a2b0d3c5b9d10377926ce926087215d4e36762a8c11372d05fc82ca33153268" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.279588 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.313345 4871 scope.go:117] "RemoveContainer" containerID="c3039b7e24a5a3966ed72efc67766c99d2c5cb25a339cf4a7714e6f8c636e1d1" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.330092 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.369191 4871 scope.go:117] "RemoveContainer" containerID="01772163177e6e5575aedbf766fcb515057393c197eccbafc84345eb36f25284" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.381158 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.396477 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:21 crc kubenswrapper[4871]: E1007 22:30:21.397528 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="ceilometer-central-agent" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.397553 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="ceilometer-central-agent" Oct 07 22:30:21 crc kubenswrapper[4871]: E1007 22:30:21.397594 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="sg-core" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.397602 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="sg-core" Oct 07 22:30:21 crc kubenswrapper[4871]: E1007 22:30:21.397639 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="ceilometer-notification-agent" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.397647 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="ceilometer-notification-agent" Oct 07 22:30:21 crc kubenswrapper[4871]: E1007 22:30:21.397701 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="proxy-httpd" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.397710 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="proxy-httpd" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.398391 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="sg-core" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.401216 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="proxy-httpd" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.401252 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="ceilometer-central-agent" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.401276 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" containerName="ceilometer-notification-agent" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.405937 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.410565 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.411693 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.421624 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.421732 4871 scope.go:117] "RemoveContainer" containerID="0f840f713a6e4e09878020dc4aa7e2d0a2fa4bcf846d4e1c059b75acc895d51a" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.443765 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.444015 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-run-httpd\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.444062 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-log-httpd\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.444199 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-scripts\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.444302 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s4zw\" (UniqueName: \"kubernetes.io/projected/779bbedd-cb13-4a76-8e07-21223f1a7c85-kube-api-access-9s4zw\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.444326 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.444386 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-config-data\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.546613 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.547010 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-run-httpd\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.547117 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-log-httpd\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.547216 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-scripts\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.547328 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s4zw\" (UniqueName: \"kubernetes.io/projected/779bbedd-cb13-4a76-8e07-21223f1a7c85-kube-api-access-9s4zw\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.547416 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.547500 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-config-data\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.547644 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-log-httpd\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.547641 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-run-httpd\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.551667 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.552456 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.553106 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-scripts\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.553584 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-config-data\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.565506 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s4zw\" (UniqueName: \"kubernetes.io/projected/779bbedd-cb13-4a76-8e07-21223f1a7c85-kube-api-access-9s4zw\") pod \"ceilometer-0\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.736368 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.959484 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-lrq4w"] Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.961395 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lrq4w" Oct 07 22:30:21 crc kubenswrapper[4871]: I1007 22:30:21.972009 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lrq4w"] Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.043350 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-9p2qr"] Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.044867 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9p2qr" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.058557 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-9p2qr"] Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.061250 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kl5k\" (UniqueName: \"kubernetes.io/projected/c694d936-bdcc-4b7f-9bcd-f4552513dc0b-kube-api-access-6kl5k\") pod \"nova-api-db-create-lrq4w\" (UID: \"c694d936-bdcc-4b7f-9bcd-f4552513dc0b\") " pod="openstack/nova-api-db-create-lrq4w" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.163179 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rtfw\" (UniqueName: \"kubernetes.io/projected/10819804-8164-4a97-9fa0-fe2346c9f4f1-kube-api-access-2rtfw\") pod \"nova-cell0-db-create-9p2qr\" (UID: \"10819804-8164-4a97-9fa0-fe2346c9f4f1\") " pod="openstack/nova-cell0-db-create-9p2qr" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.163337 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kl5k\" (UniqueName: \"kubernetes.io/projected/c694d936-bdcc-4b7f-9bcd-f4552513dc0b-kube-api-access-6kl5k\") pod \"nova-api-db-create-lrq4w\" (UID: \"c694d936-bdcc-4b7f-9bcd-f4552513dc0b\") " pod="openstack/nova-api-db-create-lrq4w" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.189061 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kl5k\" (UniqueName: \"kubernetes.io/projected/c694d936-bdcc-4b7f-9bcd-f4552513dc0b-kube-api-access-6kl5k\") pod \"nova-api-db-create-lrq4w\" (UID: \"c694d936-bdcc-4b7f-9bcd-f4552513dc0b\") " pod="openstack/nova-api-db-create-lrq4w" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.224109 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.265417 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rtfw\" (UniqueName: \"kubernetes.io/projected/10819804-8164-4a97-9fa0-fe2346c9f4f1-kube-api-access-2rtfw\") pod \"nova-cell0-db-create-9p2qr\" (UID: \"10819804-8164-4a97-9fa0-fe2346c9f4f1\") " pod="openstack/nova-cell0-db-create-9p2qr" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.289234 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lrq4w" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.293501 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rtfw\" (UniqueName: \"kubernetes.io/projected/10819804-8164-4a97-9fa0-fe2346c9f4f1-kube-api-access-2rtfw\") pod \"nova-cell0-db-create-9p2qr\" (UID: \"10819804-8164-4a97-9fa0-fe2346c9f4f1\") " pod="openstack/nova-cell0-db-create-9p2qr" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.336832 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-6bxbc"] Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.338375 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6bxbc" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.347580 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6bxbc"] Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.361946 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:22 crc kubenswrapper[4871]: W1007 22:30:22.364475 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod779bbedd_cb13_4a76_8e07_21223f1a7c85.slice/crio-a41836ea73aa68a2f8f119a5605386b0b141aca84ab3d99fbf5efcaeab8b0bc9 WatchSource:0}: Error finding container a41836ea73aa68a2f8f119a5605386b0b141aca84ab3d99fbf5efcaeab8b0bc9: Status 404 returned error can't find the container with id a41836ea73aa68a2f8f119a5605386b0b141aca84ab3d99fbf5efcaeab8b0bc9 Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.367283 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqg86\" (UniqueName: \"kubernetes.io/projected/119e9627-9048-42c7-a8a1-5e0cf3167bce-kube-api-access-pqg86\") pod \"nova-cell1-db-create-6bxbc\" (UID: \"119e9627-9048-42c7-a8a1-5e0cf3167bce\") " pod="openstack/nova-cell1-db-create-6bxbc" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.377639 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9p2qr" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.469497 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqg86\" (UniqueName: \"kubernetes.io/projected/119e9627-9048-42c7-a8a1-5e0cf3167bce-kube-api-access-pqg86\") pod \"nova-cell1-db-create-6bxbc\" (UID: \"119e9627-9048-42c7-a8a1-5e0cf3167bce\") " pod="openstack/nova-cell1-db-create-6bxbc" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.503371 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqg86\" (UniqueName: \"kubernetes.io/projected/119e9627-9048-42c7-a8a1-5e0cf3167bce-kube-api-access-pqg86\") pod \"nova-cell1-db-create-6bxbc\" (UID: \"119e9627-9048-42c7-a8a1-5e0cf3167bce\") " pod="openstack/nova-cell1-db-create-6bxbc" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.589271 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.589355 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.617637 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.620521 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.631853 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.672078 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6bxbc" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.687154 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:22 crc kubenswrapper[4871]: I1007 22:30:22.933615 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lrq4w"] Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.006194 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="551d28e0-eb30-4837-9130-d53f9ce982ac" path="/var/lib/kubelet/pods/551d28e0-eb30-4837-9130-d53f9ce982ac/volumes" Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.009205 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-9p2qr"] Oct 07 22:30:23 crc kubenswrapper[4871]: W1007 22:30:23.014473 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10819804_8164_4a97_9fa0_fe2346c9f4f1.slice/crio-7ffd44fd6288d25add5ef0c4e2eb8920118b24f6949fbc4c85b0924fa378ab00 WatchSource:0}: Error finding container 7ffd44fd6288d25add5ef0c4e2eb8920118b24f6949fbc4c85b0924fa378ab00: Status 404 returned error can't find the container with id 7ffd44fd6288d25add5ef0c4e2eb8920118b24f6949fbc4c85b0924fa378ab00 Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.227018 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6bxbc"] Oct 07 22:30:23 crc kubenswrapper[4871]: W1007 22:30:23.231171 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod119e9627_9048_42c7_a8a1_5e0cf3167bce.slice/crio-a0d552e4d0b6af7b758a5cf3486888719a2414a78546f30e45b181bcdcf7ea8d WatchSource:0}: Error finding container a0d552e4d0b6af7b758a5cf3486888719a2414a78546f30e45b181bcdcf7ea8d: Status 404 returned error can't find the container with id a0d552e4d0b6af7b758a5cf3486888719a2414a78546f30e45b181bcdcf7ea8d Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.320893 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6bxbc" event={"ID":"119e9627-9048-42c7-a8a1-5e0cf3167bce","Type":"ContainerStarted","Data":"a0d552e4d0b6af7b758a5cf3486888719a2414a78546f30e45b181bcdcf7ea8d"} Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.322485 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerStarted","Data":"a41836ea73aa68a2f8f119a5605386b0b141aca84ab3d99fbf5efcaeab8b0bc9"} Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.324329 4871 generic.go:334] "Generic (PLEG): container finished" podID="10819804-8164-4a97-9fa0-fe2346c9f4f1" containerID="4bd3d3abc247db1b0ff7aa3bc5abe1e19e8cfebb175b85960154e77696a3ac5d" exitCode=0 Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.324394 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-9p2qr" event={"ID":"10819804-8164-4a97-9fa0-fe2346c9f4f1","Type":"ContainerDied","Data":"4bd3d3abc247db1b0ff7aa3bc5abe1e19e8cfebb175b85960154e77696a3ac5d"} Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.324412 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-9p2qr" event={"ID":"10819804-8164-4a97-9fa0-fe2346c9f4f1","Type":"ContainerStarted","Data":"7ffd44fd6288d25add5ef0c4e2eb8920118b24f6949fbc4c85b0924fa378ab00"} Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.328077 4871 generic.go:334] "Generic (PLEG): container finished" podID="c694d936-bdcc-4b7f-9bcd-f4552513dc0b" containerID="286c54211026600e80566fbac1ca116b1dff1c146ef3a9a8fb05b27ac63ebc82" exitCode=0 Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.328914 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lrq4w" event={"ID":"c694d936-bdcc-4b7f-9bcd-f4552513dc0b","Type":"ContainerDied","Data":"286c54211026600e80566fbac1ca116b1dff1c146ef3a9a8fb05b27ac63ebc82"} Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.328968 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lrq4w" event={"ID":"c694d936-bdcc-4b7f-9bcd-f4552513dc0b","Type":"ContainerStarted","Data":"52e307e2127fcd9e3e5f0a29fa83e0223c070d64a234aee1ab7c4d29df2f02fd"} Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.329823 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:23 crc kubenswrapper[4871]: I1007 22:30:23.329862 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:24 crc kubenswrapper[4871]: I1007 22:30:24.345680 4871 generic.go:334] "Generic (PLEG): container finished" podID="119e9627-9048-42c7-a8a1-5e0cf3167bce" containerID="5e73453b1f98ecb0f88fa594bdc169a23d7bfd230c5e5e2791d2410183915462" exitCode=0 Oct 07 22:30:24 crc kubenswrapper[4871]: I1007 22:30:24.346083 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6bxbc" event={"ID":"119e9627-9048-42c7-a8a1-5e0cf3167bce","Type":"ContainerDied","Data":"5e73453b1f98ecb0f88fa594bdc169a23d7bfd230c5e5e2791d2410183915462"} Oct 07 22:30:24 crc kubenswrapper[4871]: I1007 22:30:24.355699 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerStarted","Data":"a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c"} Oct 07 22:30:24 crc kubenswrapper[4871]: I1007 22:30:24.904479 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lrq4w" Oct 07 22:30:24 crc kubenswrapper[4871]: I1007 22:30:24.931586 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9p2qr" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.056119 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rtfw\" (UniqueName: \"kubernetes.io/projected/10819804-8164-4a97-9fa0-fe2346c9f4f1-kube-api-access-2rtfw\") pod \"10819804-8164-4a97-9fa0-fe2346c9f4f1\" (UID: \"10819804-8164-4a97-9fa0-fe2346c9f4f1\") " Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.056412 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kl5k\" (UniqueName: \"kubernetes.io/projected/c694d936-bdcc-4b7f-9bcd-f4552513dc0b-kube-api-access-6kl5k\") pod \"c694d936-bdcc-4b7f-9bcd-f4552513dc0b\" (UID: \"c694d936-bdcc-4b7f-9bcd-f4552513dc0b\") " Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.062102 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10819804-8164-4a97-9fa0-fe2346c9f4f1-kube-api-access-2rtfw" (OuterVolumeSpecName: "kube-api-access-2rtfw") pod "10819804-8164-4a97-9fa0-fe2346c9f4f1" (UID: "10819804-8164-4a97-9fa0-fe2346c9f4f1"). InnerVolumeSpecName "kube-api-access-2rtfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.062768 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c694d936-bdcc-4b7f-9bcd-f4552513dc0b-kube-api-access-6kl5k" (OuterVolumeSpecName: "kube-api-access-6kl5k") pod "c694d936-bdcc-4b7f-9bcd-f4552513dc0b" (UID: "c694d936-bdcc-4b7f-9bcd-f4552513dc0b"). InnerVolumeSpecName "kube-api-access-6kl5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.161575 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rtfw\" (UniqueName: \"kubernetes.io/projected/10819804-8164-4a97-9fa0-fe2346c9f4f1-kube-api-access-2rtfw\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.161724 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kl5k\" (UniqueName: \"kubernetes.io/projected/c694d936-bdcc-4b7f-9bcd-f4552513dc0b-kube-api-access-6kl5k\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.369502 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lrq4w" event={"ID":"c694d936-bdcc-4b7f-9bcd-f4552513dc0b","Type":"ContainerDied","Data":"52e307e2127fcd9e3e5f0a29fa83e0223c070d64a234aee1ab7c4d29df2f02fd"} Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.369550 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52e307e2127fcd9e3e5f0a29fa83e0223c070d64a234aee1ab7c4d29df2f02fd" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.369626 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lrq4w" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.373007 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerStarted","Data":"92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926"} Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.374598 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9p2qr" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.376023 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-9p2qr" event={"ID":"10819804-8164-4a97-9fa0-fe2346c9f4f1","Type":"ContainerDied","Data":"7ffd44fd6288d25add5ef0c4e2eb8920118b24f6949fbc4c85b0924fa378ab00"} Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.376101 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ffd44fd6288d25add5ef0c4e2eb8920118b24f6949fbc4c85b0924fa378ab00" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.429165 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.429686 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.711371 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6bxbc" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.774305 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqg86\" (UniqueName: \"kubernetes.io/projected/119e9627-9048-42c7-a8a1-5e0cf3167bce-kube-api-access-pqg86\") pod \"119e9627-9048-42c7-a8a1-5e0cf3167bce\" (UID: \"119e9627-9048-42c7-a8a1-5e0cf3167bce\") " Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.783013 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/119e9627-9048-42c7-a8a1-5e0cf3167bce-kube-api-access-pqg86" (OuterVolumeSpecName: "kube-api-access-pqg86") pod "119e9627-9048-42c7-a8a1-5e0cf3167bce" (UID: "119e9627-9048-42c7-a8a1-5e0cf3167bce"). InnerVolumeSpecName "kube-api-access-pqg86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:25 crc kubenswrapper[4871]: I1007 22:30:25.877287 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqg86\" (UniqueName: \"kubernetes.io/projected/119e9627-9048-42c7-a8a1-5e0cf3167bce-kube-api-access-pqg86\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:26 crc kubenswrapper[4871]: I1007 22:30:26.385247 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6bxbc" event={"ID":"119e9627-9048-42c7-a8a1-5e0cf3167bce","Type":"ContainerDied","Data":"a0d552e4d0b6af7b758a5cf3486888719a2414a78546f30e45b181bcdcf7ea8d"} Oct 07 22:30:26 crc kubenswrapper[4871]: I1007 22:30:26.385826 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0d552e4d0b6af7b758a5cf3486888719a2414a78546f30e45b181bcdcf7ea8d" Oct 07 22:30:26 crc kubenswrapper[4871]: I1007 22:30:26.385349 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6bxbc" Oct 07 22:30:27 crc kubenswrapper[4871]: I1007 22:30:27.399782 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerStarted","Data":"0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58"} Oct 07 22:30:30 crc kubenswrapper[4871]: I1007 22:30:30.432316 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerStarted","Data":"f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7"} Oct 07 22:30:30 crc kubenswrapper[4871]: I1007 22:30:30.433004 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 22:30:30 crc kubenswrapper[4871]: I1007 22:30:30.432509 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="ceilometer-central-agent" containerID="cri-o://a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c" gracePeriod=30 Oct 07 22:30:30 crc kubenswrapper[4871]: I1007 22:30:30.433167 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="proxy-httpd" containerID="cri-o://f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7" gracePeriod=30 Oct 07 22:30:30 crc kubenswrapper[4871]: I1007 22:30:30.433270 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="ceilometer-notification-agent" containerID="cri-o://92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926" gracePeriod=30 Oct 07 22:30:30 crc kubenswrapper[4871]: I1007 22:30:30.433287 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="sg-core" containerID="cri-o://0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58" gracePeriod=30 Oct 07 22:30:30 crc kubenswrapper[4871]: I1007 22:30:30.471724 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.529566622 podStartE2EDuration="9.471706397s" podCreationTimestamp="2025-10-07 22:30:21 +0000 UTC" firstStartedPulling="2025-10-07 22:30:22.372353718 +0000 UTC m=+1296.175051791" lastFinishedPulling="2025-10-07 22:30:29.314493493 +0000 UTC m=+1303.117191566" observedRunningTime="2025-10-07 22:30:30.462962096 +0000 UTC m=+1304.265660169" watchObservedRunningTime="2025-10-07 22:30:30.471706397 +0000 UTC m=+1304.274404470" Oct 07 22:30:31 crc kubenswrapper[4871]: I1007 22:30:31.451994 4871 generic.go:334] "Generic (PLEG): container finished" podID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerID="f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7" exitCode=0 Oct 07 22:30:31 crc kubenswrapper[4871]: I1007 22:30:31.452440 4871 generic.go:334] "Generic (PLEG): container finished" podID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerID="0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58" exitCode=2 Oct 07 22:30:31 crc kubenswrapper[4871]: I1007 22:30:31.452064 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerDied","Data":"f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7"} Oct 07 22:30:31 crc kubenswrapper[4871]: I1007 22:30:31.452534 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerDied","Data":"0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58"} Oct 07 22:30:31 crc kubenswrapper[4871]: I1007 22:30:31.452560 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerDied","Data":"92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926"} Oct 07 22:30:31 crc kubenswrapper[4871]: I1007 22:30:31.452456 4871 generic.go:334] "Generic (PLEG): container finished" podID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerID="92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926" exitCode=0 Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.094632 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-d46f-account-create-7qnbt"] Oct 07 22:30:32 crc kubenswrapper[4871]: E1007 22:30:32.095545 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10819804-8164-4a97-9fa0-fe2346c9f4f1" containerName="mariadb-database-create" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.095660 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="10819804-8164-4a97-9fa0-fe2346c9f4f1" containerName="mariadb-database-create" Oct 07 22:30:32 crc kubenswrapper[4871]: E1007 22:30:32.095789 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c694d936-bdcc-4b7f-9bcd-f4552513dc0b" containerName="mariadb-database-create" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.095890 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c694d936-bdcc-4b7f-9bcd-f4552513dc0b" containerName="mariadb-database-create" Oct 07 22:30:32 crc kubenswrapper[4871]: E1007 22:30:32.095997 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="119e9627-9048-42c7-a8a1-5e0cf3167bce" containerName="mariadb-database-create" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.096072 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="119e9627-9048-42c7-a8a1-5e0cf3167bce" containerName="mariadb-database-create" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.096391 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="c694d936-bdcc-4b7f-9bcd-f4552513dc0b" containerName="mariadb-database-create" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.096467 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="10819804-8164-4a97-9fa0-fe2346c9f4f1" containerName="mariadb-database-create" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.096559 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="119e9627-9048-42c7-a8a1-5e0cf3167bce" containerName="mariadb-database-create" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.097438 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d46f-account-create-7qnbt" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.101294 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.111896 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d46f-account-create-7qnbt"] Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.224134 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lnnl\" (UniqueName: \"kubernetes.io/projected/bd8851e2-da40-442d-a42b-d1a5cf56bbfa-kube-api-access-4lnnl\") pod \"nova-api-d46f-account-create-7qnbt\" (UID: \"bd8851e2-da40-442d-a42b-d1a5cf56bbfa\") " pod="openstack/nova-api-d46f-account-create-7qnbt" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.322346 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-23c1-account-create-pgt7v"] Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.324657 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-23c1-account-create-pgt7v" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.326085 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lnnl\" (UniqueName: \"kubernetes.io/projected/bd8851e2-da40-442d-a42b-d1a5cf56bbfa-kube-api-access-4lnnl\") pod \"nova-api-d46f-account-create-7qnbt\" (UID: \"bd8851e2-da40-442d-a42b-d1a5cf56bbfa\") " pod="openstack/nova-api-d46f-account-create-7qnbt" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.328636 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.334486 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-23c1-account-create-pgt7v"] Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.361263 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lnnl\" (UniqueName: \"kubernetes.io/projected/bd8851e2-da40-442d-a42b-d1a5cf56bbfa-kube-api-access-4lnnl\") pod \"nova-api-d46f-account-create-7qnbt\" (UID: \"bd8851e2-da40-442d-a42b-d1a5cf56bbfa\") " pod="openstack/nova-api-d46f-account-create-7qnbt" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.428590 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98ssn\" (UniqueName: \"kubernetes.io/projected/be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d-kube-api-access-98ssn\") pod \"nova-cell0-23c1-account-create-pgt7v\" (UID: \"be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d\") " pod="openstack/nova-cell0-23c1-account-create-pgt7v" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.429288 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d46f-account-create-7qnbt" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.493099 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ae9f-account-create-pqnwn"] Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.494706 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ae9f-account-create-pqnwn" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.498760 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.510403 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ae9f-account-create-pqnwn"] Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.530885 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98ssn\" (UniqueName: \"kubernetes.io/projected/be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d-kube-api-access-98ssn\") pod \"nova-cell0-23c1-account-create-pgt7v\" (UID: \"be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d\") " pod="openstack/nova-cell0-23c1-account-create-pgt7v" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.552645 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98ssn\" (UniqueName: \"kubernetes.io/projected/be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d-kube-api-access-98ssn\") pod \"nova-cell0-23c1-account-create-pgt7v\" (UID: \"be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d\") " pod="openstack/nova-cell0-23c1-account-create-pgt7v" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.632893 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcfx8\" (UniqueName: \"kubernetes.io/projected/1191dd97-3c15-4d04-b39a-ac4eec992a75-kube-api-access-lcfx8\") pod \"nova-cell1-ae9f-account-create-pqnwn\" (UID: \"1191dd97-3c15-4d04-b39a-ac4eec992a75\") " pod="openstack/nova-cell1-ae9f-account-create-pqnwn" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.643675 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-23c1-account-create-pgt7v" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.739322 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcfx8\" (UniqueName: \"kubernetes.io/projected/1191dd97-3c15-4d04-b39a-ac4eec992a75-kube-api-access-lcfx8\") pod \"nova-cell1-ae9f-account-create-pqnwn\" (UID: \"1191dd97-3c15-4d04-b39a-ac4eec992a75\") " pod="openstack/nova-cell1-ae9f-account-create-pqnwn" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.760419 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcfx8\" (UniqueName: \"kubernetes.io/projected/1191dd97-3c15-4d04-b39a-ac4eec992a75-kube-api-access-lcfx8\") pod \"nova-cell1-ae9f-account-create-pqnwn\" (UID: \"1191dd97-3c15-4d04-b39a-ac4eec992a75\") " pod="openstack/nova-cell1-ae9f-account-create-pqnwn" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.927292 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ae9f-account-create-pqnwn" Oct 07 22:30:32 crc kubenswrapper[4871]: I1007 22:30:32.947228 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d46f-account-create-7qnbt"] Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.125588 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-23c1-account-create-pgt7v"] Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.507189 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.521067 4871 generic.go:334] "Generic (PLEG): container finished" podID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerID="a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c" exitCode=0 Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.521146 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerDied","Data":"a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c"} Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.521179 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779bbedd-cb13-4a76-8e07-21223f1a7c85","Type":"ContainerDied","Data":"a41836ea73aa68a2f8f119a5605386b0b141aca84ab3d99fbf5efcaeab8b0bc9"} Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.521199 4871 scope.go:117] "RemoveContainer" containerID="f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.521358 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.525492 4871 generic.go:334] "Generic (PLEG): container finished" podID="be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d" containerID="ee44fec5c9a0421e422d1f7b00f0a817126ddfb1c416825eb1642a57c8becf28" exitCode=0 Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.525572 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-23c1-account-create-pgt7v" event={"ID":"be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d","Type":"ContainerDied","Data":"ee44fec5c9a0421e422d1f7b00f0a817126ddfb1c416825eb1642a57c8becf28"} Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.525604 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-23c1-account-create-pgt7v" event={"ID":"be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d","Type":"ContainerStarted","Data":"d1f428c387c2ac7ebb56e8a16b822febc66f59bf0dd5f63427c20f807fe2c6de"} Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.527713 4871 generic.go:334] "Generic (PLEG): container finished" podID="bd8851e2-da40-442d-a42b-d1a5cf56bbfa" containerID="d4a49dd5137df3e1d29d45f30f2df5cef25e935b8d3ca458065782634e5b1f86" exitCode=0 Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.527757 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d46f-account-create-7qnbt" event={"ID":"bd8851e2-da40-442d-a42b-d1a5cf56bbfa","Type":"ContainerDied","Data":"d4a49dd5137df3e1d29d45f30f2df5cef25e935b8d3ca458065782634e5b1f86"} Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.527851 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d46f-account-create-7qnbt" event={"ID":"bd8851e2-da40-442d-a42b-d1a5cf56bbfa","Type":"ContainerStarted","Data":"fa3b1223a7fd1b887e62e22726037f0cd2e7b96e5a0f640f7011bab0c16dc2bf"} Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.619627 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ae9f-account-create-pqnwn"] Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.626832 4871 scope.go:117] "RemoveContainer" containerID="0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.658655 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-log-httpd\") pod \"779bbedd-cb13-4a76-8e07-21223f1a7c85\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.658731 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-run-httpd\") pod \"779bbedd-cb13-4a76-8e07-21223f1a7c85\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.658874 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-config-data\") pod \"779bbedd-cb13-4a76-8e07-21223f1a7c85\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.658975 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s4zw\" (UniqueName: \"kubernetes.io/projected/779bbedd-cb13-4a76-8e07-21223f1a7c85-kube-api-access-9s4zw\") pod \"779bbedd-cb13-4a76-8e07-21223f1a7c85\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.659049 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-sg-core-conf-yaml\") pod \"779bbedd-cb13-4a76-8e07-21223f1a7c85\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.659083 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-combined-ca-bundle\") pod \"779bbedd-cb13-4a76-8e07-21223f1a7c85\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.659132 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-scripts\") pod \"779bbedd-cb13-4a76-8e07-21223f1a7c85\" (UID: \"779bbedd-cb13-4a76-8e07-21223f1a7c85\") " Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.660818 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "779bbedd-cb13-4a76-8e07-21223f1a7c85" (UID: "779bbedd-cb13-4a76-8e07-21223f1a7c85"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.660908 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "779bbedd-cb13-4a76-8e07-21223f1a7c85" (UID: "779bbedd-cb13-4a76-8e07-21223f1a7c85"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.667650 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-scripts" (OuterVolumeSpecName: "scripts") pod "779bbedd-cb13-4a76-8e07-21223f1a7c85" (UID: "779bbedd-cb13-4a76-8e07-21223f1a7c85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.668217 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/779bbedd-cb13-4a76-8e07-21223f1a7c85-kube-api-access-9s4zw" (OuterVolumeSpecName: "kube-api-access-9s4zw") pod "779bbedd-cb13-4a76-8e07-21223f1a7c85" (UID: "779bbedd-cb13-4a76-8e07-21223f1a7c85"). InnerVolumeSpecName "kube-api-access-9s4zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.718732 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "779bbedd-cb13-4a76-8e07-21223f1a7c85" (UID: "779bbedd-cb13-4a76-8e07-21223f1a7c85"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.761749 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.761837 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.761848 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.761855 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779bbedd-cb13-4a76-8e07-21223f1a7c85-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.761864 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s4zw\" (UniqueName: \"kubernetes.io/projected/779bbedd-cb13-4a76-8e07-21223f1a7c85-kube-api-access-9s4zw\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.767167 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "779bbedd-cb13-4a76-8e07-21223f1a7c85" (UID: "779bbedd-cb13-4a76-8e07-21223f1a7c85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.792476 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-config-data" (OuterVolumeSpecName: "config-data") pod "779bbedd-cb13-4a76-8e07-21223f1a7c85" (UID: "779bbedd-cb13-4a76-8e07-21223f1a7c85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.863871 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.863913 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779bbedd-cb13-4a76-8e07-21223f1a7c85-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.865428 4871 scope.go:117] "RemoveContainer" containerID="92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.906640 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.913062 4871 scope.go:117] "RemoveContainer" containerID="a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.918850 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.944990 4871 scope.go:117] "RemoveContainer" containerID="f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7" Oct 07 22:30:33 crc kubenswrapper[4871]: E1007 22:30:33.946495 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7\": container with ID starting with f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7 not found: ID does not exist" containerID="f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.946557 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7"} err="failed to get container status \"f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7\": rpc error: code = NotFound desc = could not find container \"f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7\": container with ID starting with f7dae5942a11bd3bdcdfb6b266eee798195343ff0d0de68d655b905f450df6f7 not found: ID does not exist" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.946598 4871 scope.go:117] "RemoveContainer" containerID="0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.946650 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:33 crc kubenswrapper[4871]: E1007 22:30:33.947127 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="sg-core" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947149 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="sg-core" Oct 07 22:30:33 crc kubenswrapper[4871]: E1007 22:30:33.947173 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="ceilometer-notification-agent" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947180 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="ceilometer-notification-agent" Oct 07 22:30:33 crc kubenswrapper[4871]: E1007 22:30:33.947203 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="proxy-httpd" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947211 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="proxy-httpd" Oct 07 22:30:33 crc kubenswrapper[4871]: E1007 22:30:33.947123 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58\": container with ID starting with 0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58 not found: ID does not exist" containerID="0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58" Oct 07 22:30:33 crc kubenswrapper[4871]: E1007 22:30:33.947224 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="ceilometer-central-agent" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947231 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="ceilometer-central-agent" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947268 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58"} err="failed to get container status \"0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58\": rpc error: code = NotFound desc = could not find container \"0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58\": container with ID starting with 0c56d8b8ee5e7dd178842ff5870af7af48e4b886263b723bd2be382434f49d58 not found: ID does not exist" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947331 4871 scope.go:117] "RemoveContainer" containerID="92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947424 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="sg-core" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947437 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="ceilometer-central-agent" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947450 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="proxy-httpd" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.947466 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" containerName="ceilometer-notification-agent" Oct 07 22:30:33 crc kubenswrapper[4871]: E1007 22:30:33.948116 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926\": container with ID starting with 92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926 not found: ID does not exist" containerID="92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.948168 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926"} err="failed to get container status \"92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926\": rpc error: code = NotFound desc = could not find container \"92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926\": container with ID starting with 92e6c483485624a1588c1f0dbc23d7bbd4ef9e1c3aa3e80d94f38e74939b2926 not found: ID does not exist" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.948209 4871 scope.go:117] "RemoveContainer" containerID="a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c" Oct 07 22:30:33 crc kubenswrapper[4871]: E1007 22:30:33.950319 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c\": container with ID starting with a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c not found: ID does not exist" containerID="a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.950374 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c"} err="failed to get container status \"a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c\": rpc error: code = NotFound desc = could not find container \"a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c\": container with ID starting with a8a57fe151838e3d59dc0651ae7335799282f45607e09107177c98966d0d420c not found: ID does not exist" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.953979 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.957368 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.957998 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 22:30:33 crc kubenswrapper[4871]: I1007 22:30:33.963105 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.067636 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-config-data\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.067705 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csw4p\" (UniqueName: \"kubernetes.io/projected/76a34237-60e6-40f6-97f9-d3f40584cd4a-kube-api-access-csw4p\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.067726 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.067748 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.068134 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-log-httpd\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.068705 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-scripts\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.068965 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-run-httpd\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.171545 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-config-data\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.171662 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csw4p\" (UniqueName: \"kubernetes.io/projected/76a34237-60e6-40f6-97f9-d3f40584cd4a-kube-api-access-csw4p\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.171741 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.171831 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.171893 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-log-httpd\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.172053 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-scripts\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.172110 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-run-httpd\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.172835 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-log-httpd\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.172921 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-run-httpd\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.179420 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.180352 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-scripts\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.181509 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.182756 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-config-data\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.192944 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csw4p\" (UniqueName: \"kubernetes.io/projected/76a34237-60e6-40f6-97f9-d3f40584cd4a-kube-api-access-csw4p\") pod \"ceilometer-0\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.274076 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.539153 4871 generic.go:334] "Generic (PLEG): container finished" podID="1191dd97-3c15-4d04-b39a-ac4eec992a75" containerID="3172ad3a17b669b3f933bc0df5ba7c55bf17064fd77c9124684cfa839ca3bd73" exitCode=0 Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.539221 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ae9f-account-create-pqnwn" event={"ID":"1191dd97-3c15-4d04-b39a-ac4eec992a75","Type":"ContainerDied","Data":"3172ad3a17b669b3f933bc0df5ba7c55bf17064fd77c9124684cfa839ca3bd73"} Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.539574 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ae9f-account-create-pqnwn" event={"ID":"1191dd97-3c15-4d04-b39a-ac4eec992a75","Type":"ContainerStarted","Data":"6270ef02353d256eb468d225589ad300af9b29cab7084ce642e3437de3fd59b2"} Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.794697 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:34 crc kubenswrapper[4871]: I1007 22:30:34.995604 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="779bbedd-cb13-4a76-8e07-21223f1a7c85" path="/var/lib/kubelet/pods/779bbedd-cb13-4a76-8e07-21223f1a7c85/volumes" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.053739 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d46f-account-create-7qnbt" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.061653 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-23c1-account-create-pgt7v" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.199600 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98ssn\" (UniqueName: \"kubernetes.io/projected/be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d-kube-api-access-98ssn\") pod \"be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d\" (UID: \"be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d\") " Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.199652 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lnnl\" (UniqueName: \"kubernetes.io/projected/bd8851e2-da40-442d-a42b-d1a5cf56bbfa-kube-api-access-4lnnl\") pod \"bd8851e2-da40-442d-a42b-d1a5cf56bbfa\" (UID: \"bd8851e2-da40-442d-a42b-d1a5cf56bbfa\") " Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.207377 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd8851e2-da40-442d-a42b-d1a5cf56bbfa-kube-api-access-4lnnl" (OuterVolumeSpecName: "kube-api-access-4lnnl") pod "bd8851e2-da40-442d-a42b-d1a5cf56bbfa" (UID: "bd8851e2-da40-442d-a42b-d1a5cf56bbfa"). InnerVolumeSpecName "kube-api-access-4lnnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.207479 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d-kube-api-access-98ssn" (OuterVolumeSpecName: "kube-api-access-98ssn") pod "be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d" (UID: "be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d"). InnerVolumeSpecName "kube-api-access-98ssn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.302070 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98ssn\" (UniqueName: \"kubernetes.io/projected/be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d-kube-api-access-98ssn\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.302401 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lnnl\" (UniqueName: \"kubernetes.io/projected/bd8851e2-da40-442d-a42b-d1a5cf56bbfa-kube-api-access-4lnnl\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.512831 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.512948 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.544490 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.560830 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d46f-account-create-7qnbt" event={"ID":"bd8851e2-da40-442d-a42b-d1a5cf56bbfa","Type":"ContainerDied","Data":"fa3b1223a7fd1b887e62e22726037f0cd2e7b96e5a0f640f7011bab0c16dc2bf"} Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.560896 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa3b1223a7fd1b887e62e22726037f0cd2e7b96e5a0f640f7011bab0c16dc2bf" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.560917 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d46f-account-create-7qnbt" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.563466 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerStarted","Data":"ff6d062c9923124d57018716a33bd75cbef8c0f52f535a074288d69654e1591b"} Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.566355 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-23c1-account-create-pgt7v" Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.566348 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-23c1-account-create-pgt7v" event={"ID":"be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d","Type":"ContainerDied","Data":"d1f428c387c2ac7ebb56e8a16b822febc66f59bf0dd5f63427c20f807fe2c6de"} Oct 07 22:30:35 crc kubenswrapper[4871]: I1007 22:30:35.566438 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1f428c387c2ac7ebb56e8a16b822febc66f59bf0dd5f63427c20f807fe2c6de" Oct 07 22:30:36 crc kubenswrapper[4871]: I1007 22:30:36.025251 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ae9f-account-create-pqnwn" Oct 07 22:30:36 crc kubenswrapper[4871]: I1007 22:30:36.125034 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcfx8\" (UniqueName: \"kubernetes.io/projected/1191dd97-3c15-4d04-b39a-ac4eec992a75-kube-api-access-lcfx8\") pod \"1191dd97-3c15-4d04-b39a-ac4eec992a75\" (UID: \"1191dd97-3c15-4d04-b39a-ac4eec992a75\") " Oct 07 22:30:36 crc kubenswrapper[4871]: I1007 22:30:36.140052 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1191dd97-3c15-4d04-b39a-ac4eec992a75-kube-api-access-lcfx8" (OuterVolumeSpecName: "kube-api-access-lcfx8") pod "1191dd97-3c15-4d04-b39a-ac4eec992a75" (UID: "1191dd97-3c15-4d04-b39a-ac4eec992a75"). InnerVolumeSpecName "kube-api-access-lcfx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:36 crc kubenswrapper[4871]: I1007 22:30:36.228161 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcfx8\" (UniqueName: \"kubernetes.io/projected/1191dd97-3c15-4d04-b39a-ac4eec992a75-kube-api-access-lcfx8\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:36 crc kubenswrapper[4871]: I1007 22:30:36.582886 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerStarted","Data":"10420ea8471a938e137cadf38c949388071cb87f2982f34e7004fb0dce6e46bd"} Oct 07 22:30:36 crc kubenswrapper[4871]: I1007 22:30:36.585394 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ae9f-account-create-pqnwn" event={"ID":"1191dd97-3c15-4d04-b39a-ac4eec992a75","Type":"ContainerDied","Data":"6270ef02353d256eb468d225589ad300af9b29cab7084ce642e3437de3fd59b2"} Oct 07 22:30:36 crc kubenswrapper[4871]: I1007 22:30:36.585468 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6270ef02353d256eb468d225589ad300af9b29cab7084ce642e3437de3fd59b2" Oct 07 22:30:36 crc kubenswrapper[4871]: I1007 22:30:36.585552 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ae9f-account-create-pqnwn" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.597124 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerStarted","Data":"211ed7fc24cb45b0500406441bec8d1dc13a5d2fc426a0d64122650d7dc3dd6c"} Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.598467 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4vtwx"] Oct 07 22:30:37 crc kubenswrapper[4871]: E1007 22:30:37.600411 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1191dd97-3c15-4d04-b39a-ac4eec992a75" containerName="mariadb-account-create" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.600450 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="1191dd97-3c15-4d04-b39a-ac4eec992a75" containerName="mariadb-account-create" Oct 07 22:30:37 crc kubenswrapper[4871]: E1007 22:30:37.600505 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd8851e2-da40-442d-a42b-d1a5cf56bbfa" containerName="mariadb-account-create" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.600518 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd8851e2-da40-442d-a42b-d1a5cf56bbfa" containerName="mariadb-account-create" Oct 07 22:30:37 crc kubenswrapper[4871]: E1007 22:30:37.600538 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d" containerName="mariadb-account-create" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.600549 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d" containerName="mariadb-account-create" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.600906 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="1191dd97-3c15-4d04-b39a-ac4eec992a75" containerName="mariadb-account-create" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.600938 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd8851e2-da40-442d-a42b-d1a5cf56bbfa" containerName="mariadb-account-create" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.600978 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d" containerName="mariadb-account-create" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.602023 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.606075 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.606286 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.608122 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2mqjs" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.612596 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4vtwx"] Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.779666 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-config-data\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.779855 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.779888 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-scripts\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.779960 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st2xd\" (UniqueName: \"kubernetes.io/projected/8ee6fca7-2686-4de8-8b35-d00684e49fbc-kube-api-access-st2xd\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.881897 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-config-data\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.882062 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.882106 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-scripts\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.882187 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st2xd\" (UniqueName: \"kubernetes.io/projected/8ee6fca7-2686-4de8-8b35-d00684e49fbc-kube-api-access-st2xd\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.887078 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-config-data\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.887224 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.902359 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-scripts\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.913853 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st2xd\" (UniqueName: \"kubernetes.io/projected/8ee6fca7-2686-4de8-8b35-d00684e49fbc-kube-api-access-st2xd\") pod \"nova-cell0-conductor-db-sync-4vtwx\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:37 crc kubenswrapper[4871]: I1007 22:30:37.955139 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:38 crc kubenswrapper[4871]: W1007 22:30:38.516297 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ee6fca7_2686_4de8_8b35_d00684e49fbc.slice/crio-c7c8ad959a2fa3328154c57f7dc1e869e466a7422f34f0fa228f1b586e3bd404 WatchSource:0}: Error finding container c7c8ad959a2fa3328154c57f7dc1e869e466a7422f34f0fa228f1b586e3bd404: Status 404 returned error can't find the container with id c7c8ad959a2fa3328154c57f7dc1e869e466a7422f34f0fa228f1b586e3bd404 Oct 07 22:30:38 crc kubenswrapper[4871]: I1007 22:30:38.554907 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4vtwx"] Oct 07 22:30:38 crc kubenswrapper[4871]: I1007 22:30:38.612857 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4vtwx" event={"ID":"8ee6fca7-2686-4de8-8b35-d00684e49fbc","Type":"ContainerStarted","Data":"c7c8ad959a2fa3328154c57f7dc1e869e466a7422f34f0fa228f1b586e3bd404"} Oct 07 22:30:38 crc kubenswrapper[4871]: I1007 22:30:38.614477 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerStarted","Data":"52da9b0e558e21deae2bc41363dfba17d9f926201a6328096a361e748f97d5a1"} Oct 07 22:30:40 crc kubenswrapper[4871]: I1007 22:30:40.648669 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerStarted","Data":"10c17734acca75275a835bab00124729522d83ae9108af2ad4b1d87774572056"} Oct 07 22:30:40 crc kubenswrapper[4871]: I1007 22:30:40.649610 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 22:30:40 crc kubenswrapper[4871]: I1007 22:30:40.649124 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="ceilometer-central-agent" containerID="cri-o://10420ea8471a938e137cadf38c949388071cb87f2982f34e7004fb0dce6e46bd" gracePeriod=30 Oct 07 22:30:40 crc kubenswrapper[4871]: I1007 22:30:40.649727 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="ceilometer-notification-agent" containerID="cri-o://211ed7fc24cb45b0500406441bec8d1dc13a5d2fc426a0d64122650d7dc3dd6c" gracePeriod=30 Oct 07 22:30:40 crc kubenswrapper[4871]: I1007 22:30:40.649773 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="proxy-httpd" containerID="cri-o://10c17734acca75275a835bab00124729522d83ae9108af2ad4b1d87774572056" gracePeriod=30 Oct 07 22:30:40 crc kubenswrapper[4871]: I1007 22:30:40.649732 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="sg-core" containerID="cri-o://52da9b0e558e21deae2bc41363dfba17d9f926201a6328096a361e748f97d5a1" gracePeriod=30 Oct 07 22:30:41 crc kubenswrapper[4871]: I1007 22:30:41.665110 4871 generic.go:334] "Generic (PLEG): container finished" podID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerID="10c17734acca75275a835bab00124729522d83ae9108af2ad4b1d87774572056" exitCode=0 Oct 07 22:30:41 crc kubenswrapper[4871]: I1007 22:30:41.665783 4871 generic.go:334] "Generic (PLEG): container finished" podID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerID="52da9b0e558e21deae2bc41363dfba17d9f926201a6328096a361e748f97d5a1" exitCode=2 Oct 07 22:30:41 crc kubenswrapper[4871]: I1007 22:30:41.665826 4871 generic.go:334] "Generic (PLEG): container finished" podID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerID="211ed7fc24cb45b0500406441bec8d1dc13a5d2fc426a0d64122650d7dc3dd6c" exitCode=0 Oct 07 22:30:41 crc kubenswrapper[4871]: I1007 22:30:41.665203 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerDied","Data":"10c17734acca75275a835bab00124729522d83ae9108af2ad4b1d87774572056"} Oct 07 22:30:41 crc kubenswrapper[4871]: I1007 22:30:41.665903 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerDied","Data":"52da9b0e558e21deae2bc41363dfba17d9f926201a6328096a361e748f97d5a1"} Oct 07 22:30:41 crc kubenswrapper[4871]: I1007 22:30:41.665937 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerDied","Data":"211ed7fc24cb45b0500406441bec8d1dc13a5d2fc426a0d64122650d7dc3dd6c"} Oct 07 22:30:43 crc kubenswrapper[4871]: I1007 22:30:43.688467 4871 generic.go:334] "Generic (PLEG): container finished" podID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerID="10420ea8471a938e137cadf38c949388071cb87f2982f34e7004fb0dce6e46bd" exitCode=0 Oct 07 22:30:43 crc kubenswrapper[4871]: I1007 22:30:43.688509 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerDied","Data":"10420ea8471a938e137cadf38c949388071cb87f2982f34e7004fb0dce6e46bd"} Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.205353 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.289121 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csw4p\" (UniqueName: \"kubernetes.io/projected/76a34237-60e6-40f6-97f9-d3f40584cd4a-kube-api-access-csw4p\") pod \"76a34237-60e6-40f6-97f9-d3f40584cd4a\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.289318 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-config-data\") pod \"76a34237-60e6-40f6-97f9-d3f40584cd4a\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.289398 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-run-httpd\") pod \"76a34237-60e6-40f6-97f9-d3f40584cd4a\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.289486 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-combined-ca-bundle\") pod \"76a34237-60e6-40f6-97f9-d3f40584cd4a\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.289521 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-sg-core-conf-yaml\") pod \"76a34237-60e6-40f6-97f9-d3f40584cd4a\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.289557 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-log-httpd\") pod \"76a34237-60e6-40f6-97f9-d3f40584cd4a\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.289697 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-scripts\") pod \"76a34237-60e6-40f6-97f9-d3f40584cd4a\" (UID: \"76a34237-60e6-40f6-97f9-d3f40584cd4a\") " Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.291837 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "76a34237-60e6-40f6-97f9-d3f40584cd4a" (UID: "76a34237-60e6-40f6-97f9-d3f40584cd4a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.292477 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "76a34237-60e6-40f6-97f9-d3f40584cd4a" (UID: "76a34237-60e6-40f6-97f9-d3f40584cd4a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.296986 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-scripts" (OuterVolumeSpecName: "scripts") pod "76a34237-60e6-40f6-97f9-d3f40584cd4a" (UID: "76a34237-60e6-40f6-97f9-d3f40584cd4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.299560 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a34237-60e6-40f6-97f9-d3f40584cd4a-kube-api-access-csw4p" (OuterVolumeSpecName: "kube-api-access-csw4p") pod "76a34237-60e6-40f6-97f9-d3f40584cd4a" (UID: "76a34237-60e6-40f6-97f9-d3f40584cd4a"). InnerVolumeSpecName "kube-api-access-csw4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.322516 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "76a34237-60e6-40f6-97f9-d3f40584cd4a" (UID: "76a34237-60e6-40f6-97f9-d3f40584cd4a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.384596 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76a34237-60e6-40f6-97f9-d3f40584cd4a" (UID: "76a34237-60e6-40f6-97f9-d3f40584cd4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.386954 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-config-data" (OuterVolumeSpecName: "config-data") pod "76a34237-60e6-40f6-97f9-d3f40584cd4a" (UID: "76a34237-60e6-40f6-97f9-d3f40584cd4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.392650 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.392689 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csw4p\" (UniqueName: \"kubernetes.io/projected/76a34237-60e6-40f6-97f9-d3f40584cd4a-kube-api-access-csw4p\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.392706 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.392724 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.392736 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.392749 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a34237-60e6-40f6-97f9-d3f40584cd4a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.392761 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a34237-60e6-40f6-97f9-d3f40584cd4a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.733023 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a34237-60e6-40f6-97f9-d3f40584cd4a","Type":"ContainerDied","Data":"ff6d062c9923124d57018716a33bd75cbef8c0f52f535a074288d69654e1591b"} Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.733707 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.733738 4871 scope.go:117] "RemoveContainer" containerID="10c17734acca75275a835bab00124729522d83ae9108af2ad4b1d87774572056" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.735601 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4vtwx" event={"ID":"8ee6fca7-2686-4de8-8b35-d00684e49fbc","Type":"ContainerStarted","Data":"1c378db964db6ca85fc9a95ee1643c1a2799b62adc25117e536465d06965954e"} Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.768592 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-4vtwx" podStartSLOduration=2.342066353 podStartE2EDuration="9.768564122s" podCreationTimestamp="2025-10-07 22:30:37 +0000 UTC" firstStartedPulling="2025-10-07 22:30:38.51861689 +0000 UTC m=+1312.321314973" lastFinishedPulling="2025-10-07 22:30:45.945114659 +0000 UTC m=+1319.747812742" observedRunningTime="2025-10-07 22:30:46.76622849 +0000 UTC m=+1320.568926673" watchObservedRunningTime="2025-10-07 22:30:46.768564122 +0000 UTC m=+1320.571262205" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.794841 4871 scope.go:117] "RemoveContainer" containerID="52da9b0e558e21deae2bc41363dfba17d9f926201a6328096a361e748f97d5a1" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.807426 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.834950 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.855892 4871 scope.go:117] "RemoveContainer" containerID="211ed7fc24cb45b0500406441bec8d1dc13a5d2fc426a0d64122650d7dc3dd6c" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.872561 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:46 crc kubenswrapper[4871]: E1007 22:30:46.873200 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="proxy-httpd" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.873236 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="proxy-httpd" Oct 07 22:30:46 crc kubenswrapper[4871]: E1007 22:30:46.873252 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="ceilometer-notification-agent" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.873263 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="ceilometer-notification-agent" Oct 07 22:30:46 crc kubenswrapper[4871]: E1007 22:30:46.873297 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="sg-core" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.873306 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="sg-core" Oct 07 22:30:46 crc kubenswrapper[4871]: E1007 22:30:46.873335 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="ceilometer-central-agent" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.873345 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="ceilometer-central-agent" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.873569 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="ceilometer-central-agent" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.873609 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="ceilometer-notification-agent" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.873634 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="sg-core" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.873645 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" containerName="proxy-httpd" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.876003 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.878365 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.884183 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.891894 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.924861 4871 scope.go:117] "RemoveContainer" containerID="10420ea8471a938e137cadf38c949388071cb87f2982f34e7004fb0dce6e46bd" Oct 07 22:30:46 crc kubenswrapper[4871]: I1007 22:30:46.995074 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76a34237-60e6-40f6-97f9-d3f40584cd4a" path="/var/lib/kubelet/pods/76a34237-60e6-40f6-97f9-d3f40584cd4a/volumes" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.017689 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-run-httpd\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.017747 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-config-data\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.017841 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.017871 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctf2l\" (UniqueName: \"kubernetes.io/projected/20721d6e-b77b-4d35-971f-99e8a37f1fa5-kube-api-access-ctf2l\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.017889 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-log-httpd\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.017915 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-scripts\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.017932 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.119941 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.120002 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctf2l\" (UniqueName: \"kubernetes.io/projected/20721d6e-b77b-4d35-971f-99e8a37f1fa5-kube-api-access-ctf2l\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.120027 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-log-httpd\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.120059 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-scripts\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.120078 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.120200 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-run-httpd\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.120235 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-config-data\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.120879 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-log-httpd\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.122949 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-run-httpd\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.125027 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.126253 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.134709 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.136753 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.137109 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-scripts\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.137619 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-config-data\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.155000 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctf2l\" (UniqueName: \"kubernetes.io/projected/20721d6e-b77b-4d35-971f-99e8a37f1fa5-kube-api-access-ctf2l\") pod \"ceilometer-0\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.243828 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:30:47 crc kubenswrapper[4871]: W1007 22:30:47.721477 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20721d6e_b77b_4d35_971f_99e8a37f1fa5.slice/crio-be59857a76984f471e64e71150165ce9c03cb4e5f3045ac2d52a78160a94001e WatchSource:0}: Error finding container be59857a76984f471e64e71150165ce9c03cb4e5f3045ac2d52a78160a94001e: Status 404 returned error can't find the container with id be59857a76984f471e64e71150165ce9c03cb4e5f3045ac2d52a78160a94001e Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.725956 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:30:47 crc kubenswrapper[4871]: I1007 22:30:47.750214 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerStarted","Data":"be59857a76984f471e64e71150165ce9c03cb4e5f3045ac2d52a78160a94001e"} Oct 07 22:30:48 crc kubenswrapper[4871]: I1007 22:30:48.763816 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerStarted","Data":"cf50e0e9a6b0c3c268feb02e0a34db2f632212605bad9a0b0e4cdeed5ff8ab16"} Oct 07 22:30:50 crc kubenswrapper[4871]: I1007 22:30:50.799745 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerStarted","Data":"c8d29e036c2ddb2e3bfd03d8b72746934d04f79c0da11037f21c4bdcb1a505aa"} Oct 07 22:30:51 crc kubenswrapper[4871]: I1007 22:30:51.813281 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerStarted","Data":"122c39c3841cead6bd6f3881891aa89f2022243df2baafae2b0c839f50f7574b"} Oct 07 22:30:53 crc kubenswrapper[4871]: I1007 22:30:53.834151 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerStarted","Data":"531228b6619db232856f23bc8d56d8803d5d343b1f9945fe0ac7816a9370939f"} Oct 07 22:30:53 crc kubenswrapper[4871]: I1007 22:30:53.836722 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 22:30:53 crc kubenswrapper[4871]: I1007 22:30:53.865427 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.993461173 podStartE2EDuration="7.865405376s" podCreationTimestamp="2025-10-07 22:30:46 +0000 UTC" firstStartedPulling="2025-10-07 22:30:47.726030435 +0000 UTC m=+1321.528728508" lastFinishedPulling="2025-10-07 22:30:52.597974628 +0000 UTC m=+1326.400672711" observedRunningTime="2025-10-07 22:30:53.862807988 +0000 UTC m=+1327.665506051" watchObservedRunningTime="2025-10-07 22:30:53.865405376 +0000 UTC m=+1327.668103449" Oct 07 22:30:56 crc kubenswrapper[4871]: I1007 22:30:56.895282 4871 generic.go:334] "Generic (PLEG): container finished" podID="8ee6fca7-2686-4de8-8b35-d00684e49fbc" containerID="1c378db964db6ca85fc9a95ee1643c1a2799b62adc25117e536465d06965954e" exitCode=0 Oct 07 22:30:56 crc kubenswrapper[4871]: I1007 22:30:56.895656 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4vtwx" event={"ID":"8ee6fca7-2686-4de8-8b35-d00684e49fbc","Type":"ContainerDied","Data":"1c378db964db6ca85fc9a95ee1643c1a2799b62adc25117e536465d06965954e"} Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.484196 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.614176 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st2xd\" (UniqueName: \"kubernetes.io/projected/8ee6fca7-2686-4de8-8b35-d00684e49fbc-kube-api-access-st2xd\") pod \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.614370 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-scripts\") pod \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.614412 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-combined-ca-bundle\") pod \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.616440 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-config-data\") pod \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\" (UID: \"8ee6fca7-2686-4de8-8b35-d00684e49fbc\") " Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.622264 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ee6fca7-2686-4de8-8b35-d00684e49fbc-kube-api-access-st2xd" (OuterVolumeSpecName: "kube-api-access-st2xd") pod "8ee6fca7-2686-4de8-8b35-d00684e49fbc" (UID: "8ee6fca7-2686-4de8-8b35-d00684e49fbc"). InnerVolumeSpecName "kube-api-access-st2xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.626236 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-scripts" (OuterVolumeSpecName: "scripts") pod "8ee6fca7-2686-4de8-8b35-d00684e49fbc" (UID: "8ee6fca7-2686-4de8-8b35-d00684e49fbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.645131 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-config-data" (OuterVolumeSpecName: "config-data") pod "8ee6fca7-2686-4de8-8b35-d00684e49fbc" (UID: "8ee6fca7-2686-4de8-8b35-d00684e49fbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.646295 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ee6fca7-2686-4de8-8b35-d00684e49fbc" (UID: "8ee6fca7-2686-4de8-8b35-d00684e49fbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.719591 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st2xd\" (UniqueName: \"kubernetes.io/projected/8ee6fca7-2686-4de8-8b35-d00684e49fbc-kube-api-access-st2xd\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.719638 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.719650 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.719696 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6fca7-2686-4de8-8b35-d00684e49fbc-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.949444 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4vtwx" event={"ID":"8ee6fca7-2686-4de8-8b35-d00684e49fbc","Type":"ContainerDied","Data":"c7c8ad959a2fa3328154c57f7dc1e869e466a7422f34f0fa228f1b586e3bd404"} Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.949497 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7c8ad959a2fa3328154c57f7dc1e869e466a7422f34f0fa228f1b586e3bd404" Oct 07 22:30:58 crc kubenswrapper[4871]: I1007 22:30:58.949589 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4vtwx" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.051626 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 22:30:59 crc kubenswrapper[4871]: E1007 22:30:59.052754 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee6fca7-2686-4de8-8b35-d00684e49fbc" containerName="nova-cell0-conductor-db-sync" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.052778 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee6fca7-2686-4de8-8b35-d00684e49fbc" containerName="nova-cell0-conductor-db-sync" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.053097 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ee6fca7-2686-4de8-8b35-d00684e49fbc" containerName="nova-cell0-conductor-db-sync" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.054039 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.057121 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2mqjs" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.058032 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.061940 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.136187 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.136303 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.136403 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78gk8\" (UniqueName: \"kubernetes.io/projected/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-kube-api-access-78gk8\") pod \"nova-cell0-conductor-0\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.238762 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.238935 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.239029 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78gk8\" (UniqueName: \"kubernetes.io/projected/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-kube-api-access-78gk8\") pod \"nova-cell0-conductor-0\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.246291 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.250811 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.264122 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78gk8\" (UniqueName: \"kubernetes.io/projected/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-kube-api-access-78gk8\") pod \"nova-cell0-conductor-0\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " pod="openstack/nova-cell0-conductor-0" Oct 07 22:30:59 crc kubenswrapper[4871]: I1007 22:30:59.390674 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 22:31:00 crc kubenswrapper[4871]: I1007 22:31:00.065585 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 22:31:00 crc kubenswrapper[4871]: I1007 22:31:00.970727 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1e61b0b6-83cc-48b0-9f3d-b9b607faa830","Type":"ContainerStarted","Data":"655d7d7e1b7fb65057965cba90062d42b1f499d86be0b09af648ef03eef592b5"} Oct 07 22:31:00 crc kubenswrapper[4871]: I1007 22:31:00.971428 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 07 22:31:00 crc kubenswrapper[4871]: I1007 22:31:00.971456 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1e61b0b6-83cc-48b0-9f3d-b9b607faa830","Type":"ContainerStarted","Data":"35140a7bd0b66f48c5c839c75acf092a674587aee56df1429be08e5ab25f5fce"} Oct 07 22:31:01 crc kubenswrapper[4871]: I1007 22:31:01.003593 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.00356322 podStartE2EDuration="2.00356322s" podCreationTimestamp="2025-10-07 22:30:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:00.994695736 +0000 UTC m=+1334.797393849" watchObservedRunningTime="2025-10-07 22:31:01.00356322 +0000 UTC m=+1334.806261333" Oct 07 22:31:05 crc kubenswrapper[4871]: I1007 22:31:05.512072 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:31:05 crc kubenswrapper[4871]: I1007 22:31:05.512771 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:31:05 crc kubenswrapper[4871]: I1007 22:31:05.512875 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:31:05 crc kubenswrapper[4871]: I1007 22:31:05.514113 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a21e914393b384baa70031a146416b2ffe44375e4ba7b43b6aad9a92b38797c6"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:31:05 crc kubenswrapper[4871]: I1007 22:31:05.514234 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://a21e914393b384baa70031a146416b2ffe44375e4ba7b43b6aad9a92b38797c6" gracePeriod=600 Oct 07 22:31:06 crc kubenswrapper[4871]: I1007 22:31:06.036668 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="a21e914393b384baa70031a146416b2ffe44375e4ba7b43b6aad9a92b38797c6" exitCode=0 Oct 07 22:31:06 crc kubenswrapper[4871]: I1007 22:31:06.036767 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"a21e914393b384baa70031a146416b2ffe44375e4ba7b43b6aad9a92b38797c6"} Oct 07 22:31:06 crc kubenswrapper[4871]: I1007 22:31:06.037212 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6"} Oct 07 22:31:06 crc kubenswrapper[4871]: I1007 22:31:06.037252 4871 scope.go:117] "RemoveContainer" containerID="e0ffdee3d72686448819835e9ee3e3402bc2e2e3b339676f30e5a058e9fac0be" Oct 07 22:31:09 crc kubenswrapper[4871]: I1007 22:31:09.441467 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 07 22:31:09 crc kubenswrapper[4871]: I1007 22:31:09.983496 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-gz6cr"] Oct 07 22:31:09 crc kubenswrapper[4871]: I1007 22:31:09.986208 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:09 crc kubenswrapper[4871]: I1007 22:31:09.990865 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 07 22:31:09 crc kubenswrapper[4871]: I1007 22:31:09.991302 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.022015 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gz6cr"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.085361 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-scripts\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.085891 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.085972 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-config-data\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.086118 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swd9m\" (UniqueName: \"kubernetes.io/projected/945eb664-b114-4ff5-bed5-cfdaa9a99939-kube-api-access-swd9m\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.188904 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.188964 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-config-data\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.189027 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swd9m\" (UniqueName: \"kubernetes.io/projected/945eb664-b114-4ff5-bed5-cfdaa9a99939-kube-api-access-swd9m\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.189120 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-scripts\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.206578 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-scripts\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.206662 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.210372 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-config-data\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.234270 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swd9m\" (UniqueName: \"kubernetes.io/projected/945eb664-b114-4ff5-bed5-cfdaa9a99939-kube-api-access-swd9m\") pod \"nova-cell0-cell-mapping-gz6cr\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.296982 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.298399 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.311054 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.311963 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.358874 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.394006 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-config-data\") pod \"nova-scheduler-0\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.394087 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrww7\" (UniqueName: \"kubernetes.io/projected/3011f9bb-2847-429d-b7f0-2b785cbd36a6-kube-api-access-wrww7\") pod \"nova-scheduler-0\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.394137 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.408218 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.411083 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.415132 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.487988 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.523686 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrww7\" (UniqueName: \"kubernetes.io/projected/3011f9bb-2847-429d-b7f0-2b785cbd36a6-kube-api-access-wrww7\") pod \"nova-scheduler-0\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.524054 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.524104 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.524175 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.524228 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9d65\" (UniqueName: \"kubernetes.io/projected/74fab45c-4927-4ef6-b85e-05158f2af03e-kube-api-access-n9d65\") pod \"nova-cell1-novncproxy-0\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.524290 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-config-data\") pod \"nova-scheduler-0\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.542712 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.552372 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-config-data\") pod \"nova-scheduler-0\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.557993 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.559984 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.562319 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.570485 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrww7\" (UniqueName: \"kubernetes.io/projected/3011f9bb-2847-429d-b7f0-2b785cbd36a6-kube-api-access-wrww7\") pod \"nova-scheduler-0\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.626485 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rnrf\" (UniqueName: \"kubernetes.io/projected/dc3089ec-bde1-4111-8cc8-b7182249bd1b-kube-api-access-4rnrf\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.627815 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-config-data\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.628023 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.628237 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc3089ec-bde1-4111-8cc8-b7182249bd1b-logs\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.628360 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.628451 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.628577 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9d65\" (UniqueName: \"kubernetes.io/projected/74fab45c-4927-4ef6-b85e-05158f2af03e-kube-api-access-n9d65\") pod \"nova-cell1-novncproxy-0\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.630010 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.635461 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.654016 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.654989 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.679774 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.688145 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9d65\" (UniqueName: \"kubernetes.io/projected/74fab45c-4927-4ef6-b85e-05158f2af03e-kube-api-access-n9d65\") pod \"nova-cell1-novncproxy-0\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.693348 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.693445 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.700844 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.730109 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.730167 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-config-data\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.730224 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvm88\" (UniqueName: \"kubernetes.io/projected/e46f4139-688a-4661-80a9-eed14db57edf-kube-api-access-hvm88\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.730298 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e46f4139-688a-4661-80a9-eed14db57edf-logs\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.730318 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc3089ec-bde1-4111-8cc8-b7182249bd1b-logs\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.730345 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-config-data\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.730377 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.730425 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rnrf\" (UniqueName: \"kubernetes.io/projected/dc3089ec-bde1-4111-8cc8-b7182249bd1b-kube-api-access-4rnrf\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.732352 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc3089ec-bde1-4111-8cc8-b7182249bd1b-logs\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.743541 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.748300 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-config-data\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.758268 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rnrf\" (UniqueName: \"kubernetes.io/projected/dc3089ec-bde1-4111-8cc8-b7182249bd1b-kube-api-access-4rnrf\") pod \"nova-api-0\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.760609 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-z7nsb"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.766566 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.776690 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-z7nsb"] Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.827007 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.831843 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvm88\" (UniqueName: \"kubernetes.io/projected/e46f4139-688a-4661-80a9-eed14db57edf-kube-api-access-hvm88\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.831889 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-nb\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.831916 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-config\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.831972 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e46f4139-688a-4661-80a9-eed14db57edf-logs\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.831993 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-config-data\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.832012 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-svc\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.832050 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zspj\" (UniqueName: \"kubernetes.io/projected/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-kube-api-access-7zspj\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.832084 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-sb\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.832110 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.832131 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-swift-storage-0\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.839714 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e46f4139-688a-4661-80a9-eed14db57edf-logs\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.850807 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-config-data\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.853150 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.874957 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvm88\" (UniqueName: \"kubernetes.io/projected/e46f4139-688a-4661-80a9-eed14db57edf-kube-api-access-hvm88\") pod \"nova-metadata-0\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " pod="openstack/nova-metadata-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.944078 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-swift-storage-0\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.944666 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-nb\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.944699 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-config\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.944762 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-svc\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.944823 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zspj\" (UniqueName: \"kubernetes.io/projected/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-kube-api-access-7zspj\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.944859 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-sb\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.945763 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-sb\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.946362 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-swift-storage-0\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.946542 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.947901 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-nb\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.948652 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-config\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.949267 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-svc\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:10 crc kubenswrapper[4871]: I1007 22:31:10.979079 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zspj\" (UniqueName: \"kubernetes.io/projected/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-kube-api-access-7zspj\") pod \"dnsmasq-dns-64d8d96789-z7nsb\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.042394 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.123362 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.131547 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gz6cr"] Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.329950 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.655551 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dbxmm"] Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.657718 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.668106 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.668187 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.682860 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dbxmm"] Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.699390 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.765261 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.765313 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-scripts\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.765371 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph795\" (UniqueName: \"kubernetes.io/projected/034a9072-67f8-45cb-896a-b238766f2f16-kube-api-access-ph795\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.765406 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-config-data\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.846789 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.867867 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-config-data\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.868051 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.868080 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-scripts\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.868148 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph795\" (UniqueName: \"kubernetes.io/projected/034a9072-67f8-45cb-896a-b238766f2f16-kube-api-access-ph795\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.874361 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.874939 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-scripts\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.875341 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-config-data\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.889052 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph795\" (UniqueName: \"kubernetes.io/projected/034a9072-67f8-45cb-896a-b238766f2f16-kube-api-access-ph795\") pod \"nova-cell1-conductor-db-sync-dbxmm\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.965477 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-z7nsb"] Oct 07 22:31:11 crc kubenswrapper[4871]: W1007 22:31:11.967773 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bbe7cd5_b5d5_49ea_9955_72c1d41f3e4a.slice/crio-805dd6bbbc6336267286e4bb1f92a0839026a7bd2399b71b267ae58f9b95d410 WatchSource:0}: Error finding container 805dd6bbbc6336267286e4bb1f92a0839026a7bd2399b71b267ae58f9b95d410: Status 404 returned error can't find the container with id 805dd6bbbc6336267286e4bb1f92a0839026a7bd2399b71b267ae58f9b95d410 Oct 07 22:31:11 crc kubenswrapper[4871]: I1007 22:31:11.998998 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.035828 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:12 crc kubenswrapper[4871]: W1007 22:31:12.050168 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode46f4139_688a_4661_80a9_eed14db57edf.slice/crio-b08744827262f97629db760d2cd7c623c30ee9b093c525745df72790b2afea8d WatchSource:0}: Error finding container b08744827262f97629db760d2cd7c623c30ee9b093c525745df72790b2afea8d: Status 404 returned error can't find the container with id b08744827262f97629db760d2cd7c623c30ee9b093c525745df72790b2afea8d Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.129589 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"74fab45c-4927-4ef6-b85e-05158f2af03e","Type":"ContainerStarted","Data":"b7af9a17ddcc74916d39d4bbc91d614b59c3b9452997c1a663aad4cb5590aa06"} Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.164149 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e46f4139-688a-4661-80a9-eed14db57edf","Type":"ContainerStarted","Data":"b08744827262f97629db760d2cd7c623c30ee9b093c525745df72790b2afea8d"} Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.167471 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3011f9bb-2847-429d-b7f0-2b785cbd36a6","Type":"ContainerStarted","Data":"6f38df976a509fcbe57f96d5169d9261bca3c53eaf4de4fe547448da6a92aece"} Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.168393 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc3089ec-bde1-4111-8cc8-b7182249bd1b","Type":"ContainerStarted","Data":"a5b327657073fca9d220ab56131c1d94caaa9f28f1bb250efe6b0efec1efb5b9"} Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.173399 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" event={"ID":"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a","Type":"ContainerStarted","Data":"805dd6bbbc6336267286e4bb1f92a0839026a7bd2399b71b267ae58f9b95d410"} Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.177644 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gz6cr" event={"ID":"945eb664-b114-4ff5-bed5-cfdaa9a99939","Type":"ContainerStarted","Data":"525b8530afe70e73d50536a5be4989228b6800d6312af9fbc65f81c15ad00346"} Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.177675 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gz6cr" event={"ID":"945eb664-b114-4ff5-bed5-cfdaa9a99939","Type":"ContainerStarted","Data":"67b39330358fe794b6e8c4d50a6593c27f74aef9af1c9f887a5c040b5f7acbc8"} Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.213363 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-gz6cr" podStartSLOduration=3.213338653 podStartE2EDuration="3.213338653s" podCreationTimestamp="2025-10-07 22:31:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:12.194707893 +0000 UTC m=+1345.997405966" watchObservedRunningTime="2025-10-07 22:31:12.213338653 +0000 UTC m=+1346.016036726" Oct 07 22:31:12 crc kubenswrapper[4871]: I1007 22:31:12.622102 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dbxmm"] Oct 07 22:31:13 crc kubenswrapper[4871]: I1007 22:31:13.191586 4871 generic.go:334] "Generic (PLEG): container finished" podID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" containerID="c09d7323fe49b6e9493519e2e016807782896bb28fb648e44c167ed8fa4b538a" exitCode=0 Oct 07 22:31:13 crc kubenswrapper[4871]: I1007 22:31:13.192136 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" event={"ID":"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a","Type":"ContainerDied","Data":"c09d7323fe49b6e9493519e2e016807782896bb28fb648e44c167ed8fa4b538a"} Oct 07 22:31:13 crc kubenswrapper[4871]: I1007 22:31:13.195400 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dbxmm" event={"ID":"034a9072-67f8-45cb-896a-b238766f2f16","Type":"ContainerStarted","Data":"45785f0d795f8ab376fa609b6da3c877ac64fc1ed8340a47e9be2e197d469507"} Oct 07 22:31:13 crc kubenswrapper[4871]: I1007 22:31:13.195585 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dbxmm" event={"ID":"034a9072-67f8-45cb-896a-b238766f2f16","Type":"ContainerStarted","Data":"00cb06b9f621e78147c2996052a4fd459c50e12d2cd93e24ea9fa0e66e2c8168"} Oct 07 22:31:13 crc kubenswrapper[4871]: I1007 22:31:13.241370 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-dbxmm" podStartSLOduration=2.241330253 podStartE2EDuration="2.241330253s" podCreationTimestamp="2025-10-07 22:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:13.240239535 +0000 UTC m=+1347.042937608" watchObservedRunningTime="2025-10-07 22:31:13.241330253 +0000 UTC m=+1347.044028326" Oct 07 22:31:14 crc kubenswrapper[4871]: I1007 22:31:14.508701 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:14 crc kubenswrapper[4871]: I1007 22:31:14.534444 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.244832 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" event={"ID":"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a","Type":"ContainerStarted","Data":"ac7926263bebb46db052a48e0d59ffb573ae16e631980ef5e58f7119d8da1a49"} Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.245965 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.274156 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"74fab45c-4927-4ef6-b85e-05158f2af03e","Type":"ContainerStarted","Data":"e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49"} Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.274377 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="74fab45c-4927-4ef6-b85e-05158f2af03e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49" gracePeriod=30 Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.305139 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e46f4139-688a-4661-80a9-eed14db57edf","Type":"ContainerStarted","Data":"e096f2599046a8f696bda33001a728ac4a25cbdd627e55a169a41761325f3d89"} Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.312914 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" podStartSLOduration=6.31288711 podStartE2EDuration="6.31288711s" podCreationTimestamp="2025-10-07 22:31:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:16.305984978 +0000 UTC m=+1350.108683051" watchObservedRunningTime="2025-10-07 22:31:16.31288711 +0000 UTC m=+1350.115585183" Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.313255 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3011f9bb-2847-429d-b7f0-2b785cbd36a6","Type":"ContainerStarted","Data":"7bc28f2f93b542fa3e68c12bf8738025863351bb15b09eb735cd87e81603e187"} Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.324162 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc3089ec-bde1-4111-8cc8-b7182249bd1b","Type":"ContainerStarted","Data":"4a961d85346032be26517f73757ff23c9ee4d8889ca199b46a78f970ab005b6e"} Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.341293 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.43414936 podStartE2EDuration="6.341269087s" podCreationTimestamp="2025-10-07 22:31:10 +0000 UTC" firstStartedPulling="2025-10-07 22:31:11.685100308 +0000 UTC m=+1345.487798381" lastFinishedPulling="2025-10-07 22:31:15.592220035 +0000 UTC m=+1349.394918108" observedRunningTime="2025-10-07 22:31:16.327460613 +0000 UTC m=+1350.130158696" watchObservedRunningTime="2025-10-07 22:31:16.341269087 +0000 UTC m=+1350.143967160" Oct 07 22:31:16 crc kubenswrapper[4871]: I1007 22:31:16.367016 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.188721165 podStartE2EDuration="6.366993485s" podCreationTimestamp="2025-10-07 22:31:10 +0000 UTC" firstStartedPulling="2025-10-07 22:31:11.413140983 +0000 UTC m=+1345.215839056" lastFinishedPulling="2025-10-07 22:31:15.591413283 +0000 UTC m=+1349.394111376" observedRunningTime="2025-10-07 22:31:16.363560124 +0000 UTC m=+1350.166258197" watchObservedRunningTime="2025-10-07 22:31:16.366993485 +0000 UTC m=+1350.169691558" Oct 07 22:31:17 crc kubenswrapper[4871]: I1007 22:31:17.251180 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 22:31:17 crc kubenswrapper[4871]: I1007 22:31:17.337344 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc3089ec-bde1-4111-8cc8-b7182249bd1b","Type":"ContainerStarted","Data":"31f2ed1085d3592164382ccbfb4fe989ceda90635eece3c2661fc66104090925"} Oct 07 22:31:17 crc kubenswrapper[4871]: I1007 22:31:17.343855 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e46f4139-688a-4661-80a9-eed14db57edf","Type":"ContainerStarted","Data":"9cc45c94b48543b5d53e79a221a39f502fefbf87e40979700a4c9f3f4736e61b"} Oct 07 22:31:17 crc kubenswrapper[4871]: I1007 22:31:17.344090 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e46f4139-688a-4661-80a9-eed14db57edf" containerName="nova-metadata-log" containerID="cri-o://e096f2599046a8f696bda33001a728ac4a25cbdd627e55a169a41761325f3d89" gracePeriod=30 Oct 07 22:31:17 crc kubenswrapper[4871]: I1007 22:31:17.344218 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e46f4139-688a-4661-80a9-eed14db57edf" containerName="nova-metadata-metadata" containerID="cri-o://9cc45c94b48543b5d53e79a221a39f502fefbf87e40979700a4c9f3f4736e61b" gracePeriod=30 Oct 07 22:31:17 crc kubenswrapper[4871]: I1007 22:31:17.390966 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.643624232 podStartE2EDuration="7.390943529s" podCreationTimestamp="2025-10-07 22:31:10 +0000 UTC" firstStartedPulling="2025-10-07 22:31:11.853392611 +0000 UTC m=+1345.656090684" lastFinishedPulling="2025-10-07 22:31:15.600711918 +0000 UTC m=+1349.403409981" observedRunningTime="2025-10-07 22:31:17.364564524 +0000 UTC m=+1351.167262607" watchObservedRunningTime="2025-10-07 22:31:17.390943529 +0000 UTC m=+1351.193641602" Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.379052 4871 generic.go:334] "Generic (PLEG): container finished" podID="e46f4139-688a-4661-80a9-eed14db57edf" containerID="9cc45c94b48543b5d53e79a221a39f502fefbf87e40979700a4c9f3f4736e61b" exitCode=0 Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.379396 4871 generic.go:334] "Generic (PLEG): container finished" podID="e46f4139-688a-4661-80a9-eed14db57edf" containerID="e096f2599046a8f696bda33001a728ac4a25cbdd627e55a169a41761325f3d89" exitCode=143 Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.379147 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e46f4139-688a-4661-80a9-eed14db57edf","Type":"ContainerDied","Data":"9cc45c94b48543b5d53e79a221a39f502fefbf87e40979700a4c9f3f4736e61b"} Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.380031 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e46f4139-688a-4661-80a9-eed14db57edf","Type":"ContainerDied","Data":"e096f2599046a8f696bda33001a728ac4a25cbdd627e55a169a41761325f3d89"} Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.560122 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.646680 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvm88\" (UniqueName: \"kubernetes.io/projected/e46f4139-688a-4661-80a9-eed14db57edf-kube-api-access-hvm88\") pod \"e46f4139-688a-4661-80a9-eed14db57edf\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.647340 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e46f4139-688a-4661-80a9-eed14db57edf-logs\") pod \"e46f4139-688a-4661-80a9-eed14db57edf\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.647429 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-combined-ca-bundle\") pod \"e46f4139-688a-4661-80a9-eed14db57edf\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.647485 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-config-data\") pod \"e46f4139-688a-4661-80a9-eed14db57edf\" (UID: \"e46f4139-688a-4661-80a9-eed14db57edf\") " Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.647975 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e46f4139-688a-4661-80a9-eed14db57edf-logs" (OuterVolumeSpecName: "logs") pod "e46f4139-688a-4661-80a9-eed14db57edf" (UID: "e46f4139-688a-4661-80a9-eed14db57edf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.675244 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e46f4139-688a-4661-80a9-eed14db57edf-kube-api-access-hvm88" (OuterVolumeSpecName: "kube-api-access-hvm88") pod "e46f4139-688a-4661-80a9-eed14db57edf" (UID: "e46f4139-688a-4661-80a9-eed14db57edf"). InnerVolumeSpecName "kube-api-access-hvm88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.685403 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e46f4139-688a-4661-80a9-eed14db57edf" (UID: "e46f4139-688a-4661-80a9-eed14db57edf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.705053 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-config-data" (OuterVolumeSpecName: "config-data") pod "e46f4139-688a-4661-80a9-eed14db57edf" (UID: "e46f4139-688a-4661-80a9-eed14db57edf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.749004 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e46f4139-688a-4661-80a9-eed14db57edf-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.749040 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.749049 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e46f4139-688a-4661-80a9-eed14db57edf-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:18 crc kubenswrapper[4871]: I1007 22:31:18.749060 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvm88\" (UniqueName: \"kubernetes.io/projected/e46f4139-688a-4661-80a9-eed14db57edf-kube-api-access-hvm88\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.399563 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e46f4139-688a-4661-80a9-eed14db57edf","Type":"ContainerDied","Data":"b08744827262f97629db760d2cd7c623c30ee9b093c525745df72790b2afea8d"} Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.399631 4871 scope.go:117] "RemoveContainer" containerID="9cc45c94b48543b5d53e79a221a39f502fefbf87e40979700a4c9f3f4736e61b" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.399778 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.452883 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.466954 4871 scope.go:117] "RemoveContainer" containerID="e096f2599046a8f696bda33001a728ac4a25cbdd627e55a169a41761325f3d89" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.475878 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.486603 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:19 crc kubenswrapper[4871]: E1007 22:31:19.488407 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e46f4139-688a-4661-80a9-eed14db57edf" containerName="nova-metadata-log" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.488433 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e46f4139-688a-4661-80a9-eed14db57edf" containerName="nova-metadata-log" Oct 07 22:31:19 crc kubenswrapper[4871]: E1007 22:31:19.488461 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e46f4139-688a-4661-80a9-eed14db57edf" containerName="nova-metadata-metadata" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.488469 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e46f4139-688a-4661-80a9-eed14db57edf" containerName="nova-metadata-metadata" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.488697 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e46f4139-688a-4661-80a9-eed14db57edf" containerName="nova-metadata-metadata" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.488712 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e46f4139-688a-4661-80a9-eed14db57edf" containerName="nova-metadata-log" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.497649 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.498917 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.500419 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.500681 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.676752 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r84vv\" (UniqueName: \"kubernetes.io/projected/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-kube-api-access-r84vv\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.676958 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-config-data\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.677373 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.677620 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.677766 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-logs\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.779706 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.779771 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-logs\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.779866 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r84vv\" (UniqueName: \"kubernetes.io/projected/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-kube-api-access-r84vv\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.779908 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-config-data\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.780045 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.780953 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-logs\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.785340 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.786272 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.792855 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-config-data\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.798658 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r84vv\" (UniqueName: \"kubernetes.io/projected/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-kube-api-access-r84vv\") pod \"nova-metadata-0\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " pod="openstack/nova-metadata-0" Oct 07 22:31:19 crc kubenswrapper[4871]: I1007 22:31:19.816637 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:20 crc kubenswrapper[4871]: I1007 22:31:20.338844 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:20 crc kubenswrapper[4871]: I1007 22:31:20.419847 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb","Type":"ContainerStarted","Data":"59a6671fc9ff96a66d202340985bbd1617a029fc5d15c4afd75f8978bbee4b34"} Oct 07 22:31:20 crc kubenswrapper[4871]: I1007 22:31:20.656912 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 22:31:20 crc kubenswrapper[4871]: I1007 22:31:20.657294 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 22:31:20 crc kubenswrapper[4871]: I1007 22:31:20.702804 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 22:31:20 crc kubenswrapper[4871]: I1007 22:31:20.828005 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:20 crc kubenswrapper[4871]: I1007 22:31:20.948065 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 22:31:20 crc kubenswrapper[4871]: I1007 22:31:20.948140 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 22:31:20 crc kubenswrapper[4871]: I1007 22:31:20.995244 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e46f4139-688a-4661-80a9-eed14db57edf" path="/var/lib/kubelet/pods/e46f4139-688a-4661-80a9-eed14db57edf/volumes" Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.126075 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.206410 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-thpsg"] Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.206692 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" podUID="f73d961e-9fc8-48dd-baf3-56315f5b41a9" containerName="dnsmasq-dns" containerID="cri-o://99727673524e91dec44e277b7293baff9da10c7f31708a6b980163103ba5f999" gracePeriod=10 Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.456002 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb","Type":"ContainerStarted","Data":"222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f"} Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.456103 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb","Type":"ContainerStarted","Data":"3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209"} Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.459985 4871 generic.go:334] "Generic (PLEG): container finished" podID="034a9072-67f8-45cb-896a-b238766f2f16" containerID="45785f0d795f8ab376fa609b6da3c877ac64fc1ed8340a47e9be2e197d469507" exitCode=0 Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.460164 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dbxmm" event={"ID":"034a9072-67f8-45cb-896a-b238766f2f16","Type":"ContainerDied","Data":"45785f0d795f8ab376fa609b6da3c877ac64fc1ed8340a47e9be2e197d469507"} Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.472108 4871 generic.go:334] "Generic (PLEG): container finished" podID="945eb664-b114-4ff5-bed5-cfdaa9a99939" containerID="525b8530afe70e73d50536a5be4989228b6800d6312af9fbc65f81c15ad00346" exitCode=0 Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.472189 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gz6cr" event={"ID":"945eb664-b114-4ff5-bed5-cfdaa9a99939","Type":"ContainerDied","Data":"525b8530afe70e73d50536a5be4989228b6800d6312af9fbc65f81c15ad00346"} Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.476782 4871 generic.go:334] "Generic (PLEG): container finished" podID="f73d961e-9fc8-48dd-baf3-56315f5b41a9" containerID="99727673524e91dec44e277b7293baff9da10c7f31708a6b980163103ba5f999" exitCode=0 Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.479980 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" event={"ID":"f73d961e-9fc8-48dd-baf3-56315f5b41a9","Type":"ContainerDied","Data":"99727673524e91dec44e277b7293baff9da10c7f31708a6b980163103ba5f999"} Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.503295 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.503275881 podStartE2EDuration="2.503275881s" podCreationTimestamp="2025-10-07 22:31:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:21.501439853 +0000 UTC m=+1355.304137926" watchObservedRunningTime="2025-10-07 22:31:21.503275881 +0000 UTC m=+1355.305973954" Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.547212 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.798938 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.925454 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-config\") pod \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.926229 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-nb\") pod \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.926404 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-sb\") pod \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.926706 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxszr\" (UniqueName: \"kubernetes.io/projected/f73d961e-9fc8-48dd-baf3-56315f5b41a9-kube-api-access-jxszr\") pod \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.926756 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-svc\") pod \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.926801 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-swift-storage-0\") pod \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\" (UID: \"f73d961e-9fc8-48dd-baf3-56315f5b41a9\") " Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.936004 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f73d961e-9fc8-48dd-baf3-56315f5b41a9-kube-api-access-jxszr" (OuterVolumeSpecName: "kube-api-access-jxszr") pod "f73d961e-9fc8-48dd-baf3-56315f5b41a9" (UID: "f73d961e-9fc8-48dd-baf3-56315f5b41a9"). InnerVolumeSpecName "kube-api-access-jxszr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.993745 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f73d961e-9fc8-48dd-baf3-56315f5b41a9" (UID: "f73d961e-9fc8-48dd-baf3-56315f5b41a9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:31:21 crc kubenswrapper[4871]: I1007 22:31:21.994245 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f73d961e-9fc8-48dd-baf3-56315f5b41a9" (UID: "f73d961e-9fc8-48dd-baf3-56315f5b41a9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.002734 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-config" (OuterVolumeSpecName: "config") pod "f73d961e-9fc8-48dd-baf3-56315f5b41a9" (UID: "f73d961e-9fc8-48dd-baf3-56315f5b41a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.005442 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f73d961e-9fc8-48dd-baf3-56315f5b41a9" (UID: "f73d961e-9fc8-48dd-baf3-56315f5b41a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.029211 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxszr\" (UniqueName: \"kubernetes.io/projected/f73d961e-9fc8-48dd-baf3-56315f5b41a9-kube-api-access-jxszr\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.029266 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.029276 4871 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.029286 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.029295 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.032198 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.032308 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.036754 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f73d961e-9fc8-48dd-baf3-56315f5b41a9" (UID: "f73d961e-9fc8-48dd-baf3-56315f5b41a9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.130853 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f73d961e-9fc8-48dd-baf3-56315f5b41a9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.140748 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.141032 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="bf921c4f-b397-4a7b-b93b-b999e9cfaa8e" containerName="kube-state-metrics" containerID="cri-o://9b9a9410b344df51c50c3ea156c2ab98a9b2468099f9730d9e3e914e846bbfa3" gracePeriod=30 Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.498189 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" event={"ID":"f73d961e-9fc8-48dd-baf3-56315f5b41a9","Type":"ContainerDied","Data":"b2264180a1a26dd81bd884254bfa3c483873211f6bf0e94661a36bde5528a95a"} Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.498614 4871 scope.go:117] "RemoveContainer" containerID="99727673524e91dec44e277b7293baff9da10c7f31708a6b980163103ba5f999" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.498838 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b9656b65-thpsg" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.518722 4871 generic.go:334] "Generic (PLEG): container finished" podID="bf921c4f-b397-4a7b-b93b-b999e9cfaa8e" containerID="9b9a9410b344df51c50c3ea156c2ab98a9b2468099f9730d9e3e914e846bbfa3" exitCode=2 Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.519088 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bf921c4f-b397-4a7b-b93b-b999e9cfaa8e","Type":"ContainerDied","Data":"9b9a9410b344df51c50c3ea156c2ab98a9b2468099f9730d9e3e914e846bbfa3"} Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.546268 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-thpsg"] Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.563638 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-thpsg"] Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.614122 4871 scope.go:117] "RemoveContainer" containerID="7688829b56e07f7ed2f0e0dc39b28809ded0a114a36e436cd45d912e15025bf5" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.696362 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.850984 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9xjj\" (UniqueName: \"kubernetes.io/projected/bf921c4f-b397-4a7b-b93b-b999e9cfaa8e-kube-api-access-d9xjj\") pod \"bf921c4f-b397-4a7b-b93b-b999e9cfaa8e\" (UID: \"bf921c4f-b397-4a7b-b93b-b999e9cfaa8e\") " Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.858805 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf921c4f-b397-4a7b-b93b-b999e9cfaa8e-kube-api-access-d9xjj" (OuterVolumeSpecName: "kube-api-access-d9xjj") pod "bf921c4f-b397-4a7b-b93b-b999e9cfaa8e" (UID: "bf921c4f-b397-4a7b-b93b-b999e9cfaa8e"). InnerVolumeSpecName "kube-api-access-d9xjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:22 crc kubenswrapper[4871]: I1007 22:31:22.953850 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9xjj\" (UniqueName: \"kubernetes.io/projected/bf921c4f-b397-4a7b-b93b-b999e9cfaa8e-kube-api-access-d9xjj\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.011252 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f73d961e-9fc8-48dd-baf3-56315f5b41a9" path="/var/lib/kubelet/pods/f73d961e-9fc8-48dd-baf3-56315f5b41a9/volumes" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.105897 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.116225 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.259970 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-config-data\") pod \"945eb664-b114-4ff5-bed5-cfdaa9a99939\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.260074 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-scripts\") pod \"034a9072-67f8-45cb-896a-b238766f2f16\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.260179 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-combined-ca-bundle\") pod \"945eb664-b114-4ff5-bed5-cfdaa9a99939\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.260246 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-combined-ca-bundle\") pod \"034a9072-67f8-45cb-896a-b238766f2f16\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.260269 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-scripts\") pod \"945eb664-b114-4ff5-bed5-cfdaa9a99939\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.260360 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph795\" (UniqueName: \"kubernetes.io/projected/034a9072-67f8-45cb-896a-b238766f2f16-kube-api-access-ph795\") pod \"034a9072-67f8-45cb-896a-b238766f2f16\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.260382 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swd9m\" (UniqueName: \"kubernetes.io/projected/945eb664-b114-4ff5-bed5-cfdaa9a99939-kube-api-access-swd9m\") pod \"945eb664-b114-4ff5-bed5-cfdaa9a99939\" (UID: \"945eb664-b114-4ff5-bed5-cfdaa9a99939\") " Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.260398 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-config-data\") pod \"034a9072-67f8-45cb-896a-b238766f2f16\" (UID: \"034a9072-67f8-45cb-896a-b238766f2f16\") " Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.266375 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-scripts" (OuterVolumeSpecName: "scripts") pod "034a9072-67f8-45cb-896a-b238766f2f16" (UID: "034a9072-67f8-45cb-896a-b238766f2f16"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.266516 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-scripts" (OuterVolumeSpecName: "scripts") pod "945eb664-b114-4ff5-bed5-cfdaa9a99939" (UID: "945eb664-b114-4ff5-bed5-cfdaa9a99939"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.266623 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/945eb664-b114-4ff5-bed5-cfdaa9a99939-kube-api-access-swd9m" (OuterVolumeSpecName: "kube-api-access-swd9m") pod "945eb664-b114-4ff5-bed5-cfdaa9a99939" (UID: "945eb664-b114-4ff5-bed5-cfdaa9a99939"). InnerVolumeSpecName "kube-api-access-swd9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.267036 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/034a9072-67f8-45cb-896a-b238766f2f16-kube-api-access-ph795" (OuterVolumeSpecName: "kube-api-access-ph795") pod "034a9072-67f8-45cb-896a-b238766f2f16" (UID: "034a9072-67f8-45cb-896a-b238766f2f16"). InnerVolumeSpecName "kube-api-access-ph795". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.292893 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "945eb664-b114-4ff5-bed5-cfdaa9a99939" (UID: "945eb664-b114-4ff5-bed5-cfdaa9a99939"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.300069 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "034a9072-67f8-45cb-896a-b238766f2f16" (UID: "034a9072-67f8-45cb-896a-b238766f2f16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.312114 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-config-data" (OuterVolumeSpecName: "config-data") pod "945eb664-b114-4ff5-bed5-cfdaa9a99939" (UID: "945eb664-b114-4ff5-bed5-cfdaa9a99939"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.318000 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-config-data" (OuterVolumeSpecName: "config-data") pod "034a9072-67f8-45cb-896a-b238766f2f16" (UID: "034a9072-67f8-45cb-896a-b238766f2f16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.363303 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.363333 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.363343 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.363352 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph795\" (UniqueName: \"kubernetes.io/projected/034a9072-67f8-45cb-896a-b238766f2f16-kube-api-access-ph795\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.363361 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swd9m\" (UniqueName: \"kubernetes.io/projected/945eb664-b114-4ff5-bed5-cfdaa9a99939-kube-api-access-swd9m\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.363369 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.363379 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/945eb664-b114-4ff5-bed5-cfdaa9a99939-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.363387 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034a9072-67f8-45cb-896a-b238766f2f16-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.531446 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bf921c4f-b397-4a7b-b93b-b999e9cfaa8e","Type":"ContainerDied","Data":"fdb8ee6e311c27b3ed0db1a20a25bf2eada6c9179e05c16bbae212999cc4ed02"} Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.531516 4871 scope.go:117] "RemoveContainer" containerID="9b9a9410b344df51c50c3ea156c2ab98a9b2468099f9730d9e3e914e846bbfa3" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.531467 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.536051 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dbxmm" event={"ID":"034a9072-67f8-45cb-896a-b238766f2f16","Type":"ContainerDied","Data":"00cb06b9f621e78147c2996052a4fd459c50e12d2cd93e24ea9fa0e66e2c8168"} Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.536087 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00cb06b9f621e78147c2996052a4fd459c50e12d2cd93e24ea9fa0e66e2c8168" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.536442 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dbxmm" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.542583 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gz6cr" event={"ID":"945eb664-b114-4ff5-bed5-cfdaa9a99939","Type":"ContainerDied","Data":"67b39330358fe794b6e8c4d50a6593c27f74aef9af1c9f887a5c040b5f7acbc8"} Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.542624 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67b39330358fe794b6e8c4d50a6593c27f74aef9af1c9f887a5c040b5f7acbc8" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.542726 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gz6cr" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.576538 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.582855 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.618899 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:31:23 crc kubenswrapper[4871]: E1007 22:31:23.619548 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73d961e-9fc8-48dd-baf3-56315f5b41a9" containerName="init" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.619571 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73d961e-9fc8-48dd-baf3-56315f5b41a9" containerName="init" Oct 07 22:31:23 crc kubenswrapper[4871]: E1007 22:31:23.619583 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="945eb664-b114-4ff5-bed5-cfdaa9a99939" containerName="nova-manage" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.619589 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="945eb664-b114-4ff5-bed5-cfdaa9a99939" containerName="nova-manage" Oct 07 22:31:23 crc kubenswrapper[4871]: E1007 22:31:23.619616 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034a9072-67f8-45cb-896a-b238766f2f16" containerName="nova-cell1-conductor-db-sync" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.619622 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="034a9072-67f8-45cb-896a-b238766f2f16" containerName="nova-cell1-conductor-db-sync" Oct 07 22:31:23 crc kubenswrapper[4871]: E1007 22:31:23.619653 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73d961e-9fc8-48dd-baf3-56315f5b41a9" containerName="dnsmasq-dns" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.619662 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73d961e-9fc8-48dd-baf3-56315f5b41a9" containerName="dnsmasq-dns" Oct 07 22:31:23 crc kubenswrapper[4871]: E1007 22:31:23.619672 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf921c4f-b397-4a7b-b93b-b999e9cfaa8e" containerName="kube-state-metrics" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.619680 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf921c4f-b397-4a7b-b93b-b999e9cfaa8e" containerName="kube-state-metrics" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.619917 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf921c4f-b397-4a7b-b93b-b999e9cfaa8e" containerName="kube-state-metrics" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.619952 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f73d961e-9fc8-48dd-baf3-56315f5b41a9" containerName="dnsmasq-dns" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.619962 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="945eb664-b114-4ff5-bed5-cfdaa9a99939" containerName="nova-manage" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.619969 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="034a9072-67f8-45cb-896a-b238766f2f16" containerName="nova-cell1-conductor-db-sync" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.620734 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.635988 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.636489 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.637296 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.727310 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.731948 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.735668 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.769432 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.782231 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.782302 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.782329 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz4t2\" (UniqueName: \"kubernetes.io/projected/829bd903-9e3e-4378-9c56-6a6375d26422-kube-api-access-tz4t2\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.782381 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.825250 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.825556 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-log" containerID="cri-o://4a961d85346032be26517f73757ff23c9ee4d8889ca199b46a78f970ab005b6e" gracePeriod=30 Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.826493 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-api" containerID="cri-o://31f2ed1085d3592164382ccbfb4fe989ceda90635eece3c2661fc66104090925" gracePeriod=30 Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.850148 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.850380 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3011f9bb-2847-429d-b7f0-2b785cbd36a6" containerName="nova-scheduler-scheduler" containerID="cri-o://7bc28f2f93b542fa3e68c12bf8738025863351bb15b09eb735cd87e81603e187" gracePeriod=30 Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.884664 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.884748 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.884784 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.884828 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz4t2\" (UniqueName: \"kubernetes.io/projected/829bd903-9e3e-4378-9c56-6a6375d26422-kube-api-access-tz4t2\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.884884 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.884931 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmq8z\" (UniqueName: \"kubernetes.io/projected/fd68bd87-c8ef-4214-a9b2-299efdcb956e-kube-api-access-pmq8z\") pod \"nova-cell1-conductor-0\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.884956 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.891854 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.892476 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.893340 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.909689 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz4t2\" (UniqueName: \"kubernetes.io/projected/829bd903-9e3e-4378-9c56-6a6375d26422-kube-api-access-tz4t2\") pod \"kube-state-metrics-0\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.921731 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.924219 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerName="nova-metadata-log" containerID="cri-o://3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209" gracePeriod=30 Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.924289 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerName="nova-metadata-metadata" containerID="cri-o://222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f" gracePeriod=30 Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.959512 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.987255 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmq8z\" (UniqueName: \"kubernetes.io/projected/fd68bd87-c8ef-4214-a9b2-299efdcb956e-kube-api-access-pmq8z\") pod \"nova-cell1-conductor-0\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.987311 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.987367 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.993873 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:23 crc kubenswrapper[4871]: I1007 22:31:23.995045 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.007226 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmq8z\" (UniqueName: \"kubernetes.io/projected/fd68bd87-c8ef-4214-a9b2-299efdcb956e-kube-api-access-pmq8z\") pod \"nova-cell1-conductor-0\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.053919 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.348093 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.349236 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="sg-core" containerID="cri-o://122c39c3841cead6bd6f3881891aa89f2022243df2baafae2b0c839f50f7574b" gracePeriod=30 Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.349391 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="proxy-httpd" containerID="cri-o://531228b6619db232856f23bc8d56d8803d5d343b1f9945fe0ac7816a9370939f" gracePeriod=30 Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.349452 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="ceilometer-notification-agent" containerID="cri-o://c8d29e036c2ddb2e3bfd03d8b72746934d04f79c0da11037f21c4bdcb1a505aa" gracePeriod=30 Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.349501 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="ceilometer-central-agent" containerID="cri-o://cf50e0e9a6b0c3c268feb02e0a34db2f632212605bad9a0b0e4cdeed5ff8ab16" gracePeriod=30 Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.501842 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.558985 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.598057 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"829bd903-9e3e-4378-9c56-6a6375d26422","Type":"ContainerStarted","Data":"e2258cefc7f4d38b866c89108dd159557ee09c8dadfdfebb4aa912845fdf0cdc"} Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.607233 4871 generic.go:334] "Generic (PLEG): container finished" podID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerID="122c39c3841cead6bd6f3881891aa89f2022243df2baafae2b0c839f50f7574b" exitCode=2 Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.607348 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerDied","Data":"122c39c3841cead6bd6f3881891aa89f2022243df2baafae2b0c839f50f7574b"} Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.612252 4871 generic.go:334] "Generic (PLEG): container finished" podID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerID="4a961d85346032be26517f73757ff23c9ee4d8889ca199b46a78f970ab005b6e" exitCode=143 Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.612358 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc3089ec-bde1-4111-8cc8-b7182249bd1b","Type":"ContainerDied","Data":"4a961d85346032be26517f73757ff23c9ee4d8889ca199b46a78f970ab005b6e"} Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.614014 4871 generic.go:334] "Generic (PLEG): container finished" podID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerID="222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f" exitCode=0 Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.614038 4871 generic.go:334] "Generic (PLEG): container finished" podID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerID="3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209" exitCode=143 Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.614052 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb","Type":"ContainerDied","Data":"222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f"} Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.614070 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb","Type":"ContainerDied","Data":"3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209"} Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.614080 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb","Type":"ContainerDied","Data":"59a6671fc9ff96a66d202340985bbd1617a029fc5d15c4afd75f8978bbee4b34"} Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.614098 4871 scope.go:117] "RemoveContainer" containerID="222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.614277 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.637450 4871 scope.go:117] "RemoveContainer" containerID="3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.673025 4871 scope.go:117] "RemoveContainer" containerID="222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f" Oct 07 22:31:24 crc kubenswrapper[4871]: E1007 22:31:24.675983 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f\": container with ID starting with 222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f not found: ID does not exist" containerID="222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.676046 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f"} err="failed to get container status \"222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f\": rpc error: code = NotFound desc = could not find container \"222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f\": container with ID starting with 222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f not found: ID does not exist" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.676089 4871 scope.go:117] "RemoveContainer" containerID="3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209" Oct 07 22:31:24 crc kubenswrapper[4871]: E1007 22:31:24.676892 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209\": container with ID starting with 3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209 not found: ID does not exist" containerID="3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.676930 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209"} err="failed to get container status \"3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209\": rpc error: code = NotFound desc = could not find container \"3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209\": container with ID starting with 3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209 not found: ID does not exist" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.676961 4871 scope.go:117] "RemoveContainer" containerID="222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.677493 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f"} err="failed to get container status \"222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f\": rpc error: code = NotFound desc = could not find container \"222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f\": container with ID starting with 222d0701b5c06df55994ee9460dc0a3cde20de4d0fff8c991e6312e2f0c1795f not found: ID does not exist" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.677517 4871 scope.go:117] "RemoveContainer" containerID="3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.678130 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.678587 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209"} err="failed to get container status \"3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209\": rpc error: code = NotFound desc = could not find container \"3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209\": container with ID starting with 3975572035a38d1153f72421c5ac3be34c879f5f39947b656dd9f479624b6209 not found: ID does not exist" Oct 07 22:31:24 crc kubenswrapper[4871]: W1007 22:31:24.686526 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd68bd87_c8ef_4214_a9b2_299efdcb956e.slice/crio-cef18de0b674288b076c26646fef70938e4878c638247a3cbf2e8da910d5acd1 WatchSource:0}: Error finding container cef18de0b674288b076c26646fef70938e4878c638247a3cbf2e8da910d5acd1: Status 404 returned error can't find the container with id cef18de0b674288b076c26646fef70938e4878c638247a3cbf2e8da910d5acd1 Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.706374 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-combined-ca-bundle\") pod \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.706543 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-config-data\") pod \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.706576 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-nova-metadata-tls-certs\") pod \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.706599 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-logs\") pod \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.706751 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r84vv\" (UniqueName: \"kubernetes.io/projected/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-kube-api-access-r84vv\") pod \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\" (UID: \"79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb\") " Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.707953 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-logs" (OuterVolumeSpecName: "logs") pod "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" (UID: "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.712437 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-kube-api-access-r84vv" (OuterVolumeSpecName: "kube-api-access-r84vv") pod "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" (UID: "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb"). InnerVolumeSpecName "kube-api-access-r84vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.748426 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" (UID: "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.768430 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" (UID: "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.770766 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-config-data" (OuterVolumeSpecName: "config-data") pod "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" (UID: "79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.809948 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r84vv\" (UniqueName: \"kubernetes.io/projected/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-kube-api-access-r84vv\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.809980 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.809995 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.810003 4871 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.810012 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.967034 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.996425 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf921c4f-b397-4a7b-b93b-b999e9cfaa8e" path="/var/lib/kubelet/pods/bf921c4f-b397-4a7b-b93b-b999e9cfaa8e/volumes" Oct 07 22:31:24 crc kubenswrapper[4871]: I1007 22:31:24.998537 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.004956 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:25 crc kubenswrapper[4871]: E1007 22:31:25.040836 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerName="nova-metadata-metadata" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.040877 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerName="nova-metadata-metadata" Oct 07 22:31:25 crc kubenswrapper[4871]: E1007 22:31:25.040890 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerName="nova-metadata-log" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.040898 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerName="nova-metadata-log" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.041110 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerName="nova-metadata-metadata" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.041119 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" containerName="nova-metadata-log" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.042156 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.042267 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.044691 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.046021 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.220989 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.221345 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-config-data\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.221373 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.221393 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-logs\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.221435 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq9rk\" (UniqueName: \"kubernetes.io/projected/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-kube-api-access-kq9rk\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.323613 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.323743 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-config-data\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.323882 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.323910 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-logs\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.324182 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq9rk\" (UniqueName: \"kubernetes.io/projected/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-kube-api-access-kq9rk\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.324386 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-logs\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.328464 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.328464 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.330843 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-config-data\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.345332 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq9rk\" (UniqueName: \"kubernetes.io/projected/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-kube-api-access-kq9rk\") pod \"nova-metadata-0\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.358333 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.639249 4871 generic.go:334] "Generic (PLEG): container finished" podID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerID="531228b6619db232856f23bc8d56d8803d5d343b1f9945fe0ac7816a9370939f" exitCode=0 Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.639814 4871 generic.go:334] "Generic (PLEG): container finished" podID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerID="c8d29e036c2ddb2e3bfd03d8b72746934d04f79c0da11037f21c4bdcb1a505aa" exitCode=0 Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.639827 4871 generic.go:334] "Generic (PLEG): container finished" podID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerID="cf50e0e9a6b0c3c268feb02e0a34db2f632212605bad9a0b0e4cdeed5ff8ab16" exitCode=0 Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.639463 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerDied","Data":"531228b6619db232856f23bc8d56d8803d5d343b1f9945fe0ac7816a9370939f"} Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.639913 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerDied","Data":"c8d29e036c2ddb2e3bfd03d8b72746934d04f79c0da11037f21c4bdcb1a505aa"} Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.639926 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerDied","Data":"cf50e0e9a6b0c3c268feb02e0a34db2f632212605bad9a0b0e4cdeed5ff8ab16"} Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.643567 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"829bd903-9e3e-4378-9c56-6a6375d26422","Type":"ContainerStarted","Data":"1b18488a0daf1c9ccb071f5a1f222406f2336c9c403fbfe6cbfc92c70b53edf6"} Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.645549 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.647185 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fd68bd87-c8ef-4214-a9b2-299efdcb956e","Type":"ContainerStarted","Data":"f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63"} Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.647242 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fd68bd87-c8ef-4214-a9b2-299efdcb956e","Type":"ContainerStarted","Data":"cef18de0b674288b076c26646fef70938e4878c638247a3cbf2e8da910d5acd1"} Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.648366 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:25 crc kubenswrapper[4871]: E1007 22:31:25.658766 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7bc28f2f93b542fa3e68c12bf8738025863351bb15b09eb735cd87e81603e187" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 22:31:25 crc kubenswrapper[4871]: E1007 22:31:25.665203 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7bc28f2f93b542fa3e68c12bf8738025863351bb15b09eb735cd87e81603e187" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 22:31:25 crc kubenswrapper[4871]: E1007 22:31:25.671124 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7bc28f2f93b542fa3e68c12bf8738025863351bb15b09eb735cd87e81603e187" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 22:31:25 crc kubenswrapper[4871]: E1007 22:31:25.671235 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3011f9bb-2847-429d-b7f0-2b785cbd36a6" containerName="nova-scheduler-scheduler" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.672463 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.074700654 podStartE2EDuration="2.672434651s" podCreationTimestamp="2025-10-07 22:31:23 +0000 UTC" firstStartedPulling="2025-10-07 22:31:24.521295936 +0000 UTC m=+1358.323994009" lastFinishedPulling="2025-10-07 22:31:25.119029933 +0000 UTC m=+1358.921728006" observedRunningTime="2025-10-07 22:31:25.666382322 +0000 UTC m=+1359.469080395" watchObservedRunningTime="2025-10-07 22:31:25.672434651 +0000 UTC m=+1359.475132734" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.704011 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.703983272 podStartE2EDuration="2.703983272s" podCreationTimestamp="2025-10-07 22:31:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:25.685164567 +0000 UTC m=+1359.487862650" watchObservedRunningTime="2025-10-07 22:31:25.703983272 +0000 UTC m=+1359.506681345" Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.860358 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:31:25 crc kubenswrapper[4871]: I1007 22:31:25.887784 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.042981 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-log-httpd\") pod \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.043380 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-scripts\") pod \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.043523 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctf2l\" (UniqueName: \"kubernetes.io/projected/20721d6e-b77b-4d35-971f-99e8a37f1fa5-kube-api-access-ctf2l\") pod \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.043594 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-config-data\") pod \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.043688 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-combined-ca-bundle\") pod \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.043875 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-sg-core-conf-yaml\") pod \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.044606 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-run-httpd\") pod \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\" (UID: \"20721d6e-b77b-4d35-971f-99e8a37f1fa5\") " Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.045148 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "20721d6e-b77b-4d35-971f-99e8a37f1fa5" (UID: "20721d6e-b77b-4d35-971f-99e8a37f1fa5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.045411 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "20721d6e-b77b-4d35-971f-99e8a37f1fa5" (UID: "20721d6e-b77b-4d35-971f-99e8a37f1fa5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.046983 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.047032 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20721d6e-b77b-4d35-971f-99e8a37f1fa5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.067848 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20721d6e-b77b-4d35-971f-99e8a37f1fa5-kube-api-access-ctf2l" (OuterVolumeSpecName: "kube-api-access-ctf2l") pod "20721d6e-b77b-4d35-971f-99e8a37f1fa5" (UID: "20721d6e-b77b-4d35-971f-99e8a37f1fa5"). InnerVolumeSpecName "kube-api-access-ctf2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.069091 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-scripts" (OuterVolumeSpecName: "scripts") pod "20721d6e-b77b-4d35-971f-99e8a37f1fa5" (UID: "20721d6e-b77b-4d35-971f-99e8a37f1fa5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.088324 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "20721d6e-b77b-4d35-971f-99e8a37f1fa5" (UID: "20721d6e-b77b-4d35-971f-99e8a37f1fa5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.209232 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctf2l\" (UniqueName: \"kubernetes.io/projected/20721d6e-b77b-4d35-971f-99e8a37f1fa5-kube-api-access-ctf2l\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.209301 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.209313 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.214548 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20721d6e-b77b-4d35-971f-99e8a37f1fa5" (UID: "20721d6e-b77b-4d35-971f-99e8a37f1fa5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.237699 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-config-data" (OuterVolumeSpecName: "config-data") pod "20721d6e-b77b-4d35-971f-99e8a37f1fa5" (UID: "20721d6e-b77b-4d35-971f-99e8a37f1fa5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.310446 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.310938 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20721d6e-b77b-4d35-971f-99e8a37f1fa5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.662718 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20721d6e-b77b-4d35-971f-99e8a37f1fa5","Type":"ContainerDied","Data":"be59857a76984f471e64e71150165ce9c03cb4e5f3045ac2d52a78160a94001e"} Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.662835 4871 scope.go:117] "RemoveContainer" containerID="531228b6619db232856f23bc8d56d8803d5d343b1f9945fe0ac7816a9370939f" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.663131 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.670489 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76","Type":"ContainerStarted","Data":"be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e"} Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.670538 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76","Type":"ContainerStarted","Data":"05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101"} Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.670552 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76","Type":"ContainerStarted","Data":"5103d16c980f7aeff2b3577ef0e68425c243f15606ca5b1f36abfce357b90a51"} Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.725935 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.725897613 podStartE2EDuration="2.725897613s" podCreationTimestamp="2025-10-07 22:31:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:26.714042271 +0000 UTC m=+1360.516740384" watchObservedRunningTime="2025-10-07 22:31:26.725897613 +0000 UTC m=+1360.528595726" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.728569 4871 scope.go:117] "RemoveContainer" containerID="122c39c3841cead6bd6f3881891aa89f2022243df2baafae2b0c839f50f7574b" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.766126 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.769595 4871 scope.go:117] "RemoveContainer" containerID="c8d29e036c2ddb2e3bfd03d8b72746934d04f79c0da11037f21c4bdcb1a505aa" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.780410 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.793645 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:26 crc kubenswrapper[4871]: E1007 22:31:26.794405 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="ceilometer-central-agent" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.794442 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="ceilometer-central-agent" Oct 07 22:31:26 crc kubenswrapper[4871]: E1007 22:31:26.794462 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="ceilometer-notification-agent" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.794478 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="ceilometer-notification-agent" Oct 07 22:31:26 crc kubenswrapper[4871]: E1007 22:31:26.794499 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="sg-core" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.794513 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="sg-core" Oct 07 22:31:26 crc kubenswrapper[4871]: E1007 22:31:26.794570 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="proxy-httpd" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.794587 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="proxy-httpd" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.795024 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="proxy-httpd" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.795057 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="sg-core" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.795112 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="ceilometer-notification-agent" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.795132 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" containerName="ceilometer-central-agent" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.805630 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.812301 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.812566 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.812767 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.815811 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.822493 4871 scope.go:117] "RemoveContainer" containerID="cf50e0e9a6b0c3c268feb02e0a34db2f632212605bad9a0b0e4cdeed5ff8ab16" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.825024 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-scripts\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.826813 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.827007 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.827041 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-run-httpd\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.827082 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.827119 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-log-httpd\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.827137 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-config-data\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.827238 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x9dw\" (UniqueName: \"kubernetes.io/projected/237934fb-75da-40fe-a475-bcc78f45daa3-kube-api-access-7x9dw\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.929437 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-scripts\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.929489 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.929569 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.929591 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-run-httpd\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.929619 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.929647 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-log-httpd\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.929665 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-config-data\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.929702 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x9dw\" (UniqueName: \"kubernetes.io/projected/237934fb-75da-40fe-a475-bcc78f45daa3-kube-api-access-7x9dw\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.930498 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-run-httpd\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.930580 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-log-httpd\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.934948 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.934975 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.935488 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.936098 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-scripts\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.937721 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-config-data\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.948988 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x9dw\" (UniqueName: \"kubernetes.io/projected/237934fb-75da-40fe-a475-bcc78f45daa3-kube-api-access-7x9dw\") pod \"ceilometer-0\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " pod="openstack/ceilometer-0" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.995325 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20721d6e-b77b-4d35-971f-99e8a37f1fa5" path="/var/lib/kubelet/pods/20721d6e-b77b-4d35-971f-99e8a37f1fa5/volumes" Oct 07 22:31:26 crc kubenswrapper[4871]: I1007 22:31:26.996449 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb" path="/var/lib/kubelet/pods/79d87c4a-32bc-4a6d-9e4c-966c0b6ab3fb/volumes" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.146498 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.692439 4871 generic.go:334] "Generic (PLEG): container finished" podID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerID="31f2ed1085d3592164382ccbfb4fe989ceda90635eece3c2661fc66104090925" exitCode=0 Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.692543 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc3089ec-bde1-4111-8cc8-b7182249bd1b","Type":"ContainerDied","Data":"31f2ed1085d3592164382ccbfb4fe989ceda90635eece3c2661fc66104090925"} Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.835585 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.848293 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.869044 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-combined-ca-bundle\") pod \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.869110 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-config-data\") pod \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.869159 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rnrf\" (UniqueName: \"kubernetes.io/projected/dc3089ec-bde1-4111-8cc8-b7182249bd1b-kube-api-access-4rnrf\") pod \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.869188 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc3089ec-bde1-4111-8cc8-b7182249bd1b-logs\") pod \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\" (UID: \"dc3089ec-bde1-4111-8cc8-b7182249bd1b\") " Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.870175 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc3089ec-bde1-4111-8cc8-b7182249bd1b-logs" (OuterVolumeSpecName: "logs") pod "dc3089ec-bde1-4111-8cc8-b7182249bd1b" (UID: "dc3089ec-bde1-4111-8cc8-b7182249bd1b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.876925 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc3089ec-bde1-4111-8cc8-b7182249bd1b-kube-api-access-4rnrf" (OuterVolumeSpecName: "kube-api-access-4rnrf") pod "dc3089ec-bde1-4111-8cc8-b7182249bd1b" (UID: "dc3089ec-bde1-4111-8cc8-b7182249bd1b"). InnerVolumeSpecName "kube-api-access-4rnrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.903763 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc3089ec-bde1-4111-8cc8-b7182249bd1b" (UID: "dc3089ec-bde1-4111-8cc8-b7182249bd1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.926461 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-config-data" (OuterVolumeSpecName: "config-data") pod "dc3089ec-bde1-4111-8cc8-b7182249bd1b" (UID: "dc3089ec-bde1-4111-8cc8-b7182249bd1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.972081 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.972126 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rnrf\" (UniqueName: \"kubernetes.io/projected/dc3089ec-bde1-4111-8cc8-b7182249bd1b-kube-api-access-4rnrf\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.972140 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc3089ec-bde1-4111-8cc8-b7182249bd1b-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:27 crc kubenswrapper[4871]: I1007 22:31:27.972156 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3089ec-bde1-4111-8cc8-b7182249bd1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.712671 4871 generic.go:334] "Generic (PLEG): container finished" podID="3011f9bb-2847-429d-b7f0-2b785cbd36a6" containerID="7bc28f2f93b542fa3e68c12bf8738025863351bb15b09eb735cd87e81603e187" exitCode=0 Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.712714 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3011f9bb-2847-429d-b7f0-2b785cbd36a6","Type":"ContainerDied","Data":"7bc28f2f93b542fa3e68c12bf8738025863351bb15b09eb735cd87e81603e187"} Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.715431 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerStarted","Data":"d429507f75776fc4264a5ec8876777756d9fb21aa92148013829ca0100ac6147"} Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.719968 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc3089ec-bde1-4111-8cc8-b7182249bd1b","Type":"ContainerDied","Data":"a5b327657073fca9d220ab56131c1d94caaa9f28f1bb250efe6b0efec1efb5b9"} Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.720038 4871 scope.go:117] "RemoveContainer" containerID="31f2ed1085d3592164382ccbfb4fe989ceda90635eece3c2661fc66104090925" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.720070 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.748259 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.766506 4871 scope.go:117] "RemoveContainer" containerID="4a961d85346032be26517f73757ff23c9ee4d8889ca199b46a78f970ab005b6e" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.800277 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrww7\" (UniqueName: \"kubernetes.io/projected/3011f9bb-2847-429d-b7f0-2b785cbd36a6-kube-api-access-wrww7\") pod \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.800368 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-combined-ca-bundle\") pod \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.800442 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-config-data\") pod \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\" (UID: \"3011f9bb-2847-429d-b7f0-2b785cbd36a6\") " Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.810345 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.823512 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.836847 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:28 crc kubenswrapper[4871]: E1007 22:31:28.838400 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3011f9bb-2847-429d-b7f0-2b785cbd36a6" containerName="nova-scheduler-scheduler" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.838431 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3011f9bb-2847-429d-b7f0-2b785cbd36a6" containerName="nova-scheduler-scheduler" Oct 07 22:31:28 crc kubenswrapper[4871]: E1007 22:31:28.838470 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-api" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.838480 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-api" Oct 07 22:31:28 crc kubenswrapper[4871]: E1007 22:31:28.838546 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-log" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.838555 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-log" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.839207 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-api" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.839259 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" containerName="nova-api-log" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.839280 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3011f9bb-2847-429d-b7f0-2b785cbd36a6" containerName="nova-scheduler-scheduler" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.841972 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.844875 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.847740 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-config-data" (OuterVolumeSpecName: "config-data") pod "3011f9bb-2847-429d-b7f0-2b785cbd36a6" (UID: "3011f9bb-2847-429d-b7f0-2b785cbd36a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.848344 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3011f9bb-2847-429d-b7f0-2b785cbd36a6-kube-api-access-wrww7" (OuterVolumeSpecName: "kube-api-access-wrww7") pod "3011f9bb-2847-429d-b7f0-2b785cbd36a6" (UID: "3011f9bb-2847-429d-b7f0-2b785cbd36a6"). InnerVolumeSpecName "kube-api-access-wrww7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.879577 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.890055 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3011f9bb-2847-429d-b7f0-2b785cbd36a6" (UID: "3011f9bb-2847-429d-b7f0-2b785cbd36a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.909482 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-config-data\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.909755 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.909813 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx462\" (UniqueName: \"kubernetes.io/projected/1812a7df-9924-454b-91a7-28df2d766ab9-kube-api-access-tx462\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.911499 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1812a7df-9924-454b-91a7-28df2d766ab9-logs\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.911954 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.913061 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrww7\" (UniqueName: \"kubernetes.io/projected/3011f9bb-2847-429d-b7f0-2b785cbd36a6-kube-api-access-wrww7\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.913161 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011f9bb-2847-429d-b7f0-2b785cbd36a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:28 crc kubenswrapper[4871]: I1007 22:31:28.993287 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc3089ec-bde1-4111-8cc8-b7182249bd1b" path="/var/lib/kubelet/pods/dc3089ec-bde1-4111-8cc8-b7182249bd1b/volumes" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.014734 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.014910 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx462\" (UniqueName: \"kubernetes.io/projected/1812a7df-9924-454b-91a7-28df2d766ab9-kube-api-access-tx462\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.015078 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1812a7df-9924-454b-91a7-28df2d766ab9-logs\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.015244 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-config-data\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.017939 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1812a7df-9924-454b-91a7-28df2d766ab9-logs\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.022225 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-config-data\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.023057 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.034704 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx462\" (UniqueName: \"kubernetes.io/projected/1812a7df-9924-454b-91a7-28df2d766ab9-kube-api-access-tx462\") pod \"nova-api-0\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " pod="openstack/nova-api-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.087235 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.181990 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.641333 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:29 crc kubenswrapper[4871]: W1007 22:31:29.648970 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1812a7df_9924_454b_91a7_28df2d766ab9.slice/crio-9859c1b2fc6a984dab270551656e10fb27203831664946bdb9e32dc037e726ba WatchSource:0}: Error finding container 9859c1b2fc6a984dab270551656e10fb27203831664946bdb9e32dc037e726ba: Status 404 returned error can't find the container with id 9859c1b2fc6a984dab270551656e10fb27203831664946bdb9e32dc037e726ba Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.745671 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3011f9bb-2847-429d-b7f0-2b785cbd36a6","Type":"ContainerDied","Data":"6f38df976a509fcbe57f96d5169d9261bca3c53eaf4de4fe547448da6a92aece"} Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.745748 4871 scope.go:117] "RemoveContainer" containerID="7bc28f2f93b542fa3e68c12bf8738025863351bb15b09eb735cd87e81603e187" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.745749 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.748925 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1812a7df-9924-454b-91a7-28df2d766ab9","Type":"ContainerStarted","Data":"9859c1b2fc6a984dab270551656e10fb27203831664946bdb9e32dc037e726ba"} Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.752770 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerStarted","Data":"83982a92e0f72b913a826e96cf85b93a00c47ee62f56aeb66b806d9050b085f4"} Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.781103 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.800646 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.827079 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.830889 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.834272 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.844670 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.950876 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kkql\" (UniqueName: \"kubernetes.io/projected/a1a77ea3-6c18-41ef-a974-b285e03c0c55-kube-api-access-7kkql\") pod \"nova-scheduler-0\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.951441 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-config-data\") pod \"nova-scheduler-0\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:29 crc kubenswrapper[4871]: I1007 22:31:29.951462 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.055288 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kkql\" (UniqueName: \"kubernetes.io/projected/a1a77ea3-6c18-41ef-a974-b285e03c0c55-kube-api-access-7kkql\") pod \"nova-scheduler-0\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.055771 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.055830 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-config-data\") pod \"nova-scheduler-0\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.065475 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-config-data\") pod \"nova-scheduler-0\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.066716 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.079357 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kkql\" (UniqueName: \"kubernetes.io/projected/a1a77ea3-6c18-41ef-a974-b285e03c0c55-kube-api-access-7kkql\") pod \"nova-scheduler-0\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " pod="openstack/nova-scheduler-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.151978 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.359106 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.359550 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.430334 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4hl55"] Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.432449 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.458385 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hl55"] Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.482577 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-utilities\") pod \"redhat-marketplace-4hl55\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.483254 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98f72\" (UniqueName: \"kubernetes.io/projected/874af2ed-f68b-458a-8a82-f3c433140b9c-kube-api-access-98f72\") pod \"redhat-marketplace-4hl55\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.483422 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-catalog-content\") pod \"redhat-marketplace-4hl55\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.585913 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-catalog-content\") pod \"redhat-marketplace-4hl55\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.586057 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-utilities\") pod \"redhat-marketplace-4hl55\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.586157 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98f72\" (UniqueName: \"kubernetes.io/projected/874af2ed-f68b-458a-8a82-f3c433140b9c-kube-api-access-98f72\") pod \"redhat-marketplace-4hl55\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.587314 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-catalog-content\") pod \"redhat-marketplace-4hl55\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.587641 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-utilities\") pod \"redhat-marketplace-4hl55\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.605916 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98f72\" (UniqueName: \"kubernetes.io/projected/874af2ed-f68b-458a-8a82-f3c433140b9c-kube-api-access-98f72\") pod \"redhat-marketplace-4hl55\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.700020 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:31:30 crc kubenswrapper[4871]: W1007 22:31:30.722127 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1a77ea3_6c18_41ef_a974_b285e03c0c55.slice/crio-2ce05702e7336421aedfb550535beb0bd67cfbd9192a2cc1935c216b66f86b89 WatchSource:0}: Error finding container 2ce05702e7336421aedfb550535beb0bd67cfbd9192a2cc1935c216b66f86b89: Status 404 returned error can't find the container with id 2ce05702e7336421aedfb550535beb0bd67cfbd9192a2cc1935c216b66f86b89 Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.778175 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.796689 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1812a7df-9924-454b-91a7-28df2d766ab9","Type":"ContainerStarted","Data":"b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee"} Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.796960 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1812a7df-9924-454b-91a7-28df2d766ab9","Type":"ContainerStarted","Data":"f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa"} Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.800929 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerStarted","Data":"60d46800f21956a308b11718acac3ac7f483202813d32b1ad29888a3936484ea"} Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.804476 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a1a77ea3-6c18-41ef-a974-b285e03c0c55","Type":"ContainerStarted","Data":"2ce05702e7336421aedfb550535beb0bd67cfbd9192a2cc1935c216b66f86b89"} Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.821082 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8210607530000003 podStartE2EDuration="2.821060753s" podCreationTimestamp="2025-10-07 22:31:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:30.818844115 +0000 UTC m=+1364.621542188" watchObservedRunningTime="2025-10-07 22:31:30.821060753 +0000 UTC m=+1364.623758826" Oct 07 22:31:30 crc kubenswrapper[4871]: I1007 22:31:30.999448 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3011f9bb-2847-429d-b7f0-2b785cbd36a6" path="/var/lib/kubelet/pods/3011f9bb-2847-429d-b7f0-2b785cbd36a6/volumes" Oct 07 22:31:31 crc kubenswrapper[4871]: I1007 22:31:31.281259 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hl55"] Oct 07 22:31:31 crc kubenswrapper[4871]: W1007 22:31:31.285223 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod874af2ed_f68b_458a_8a82_f3c433140b9c.slice/crio-5c7e9b4bea4482a8096c70d1e64ddd28d86cd7427a6554892593b203724d51f8 WatchSource:0}: Error finding container 5c7e9b4bea4482a8096c70d1e64ddd28d86cd7427a6554892593b203724d51f8: Status 404 returned error can't find the container with id 5c7e9b4bea4482a8096c70d1e64ddd28d86cd7427a6554892593b203724d51f8 Oct 07 22:31:31 crc kubenswrapper[4871]: I1007 22:31:31.821377 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerStarted","Data":"1ddd5996c55ae6aba76392b6bb32551ae5d81fd8bdb6f3a5fc16898023ab622c"} Oct 07 22:31:31 crc kubenswrapper[4871]: I1007 22:31:31.823587 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a1a77ea3-6c18-41ef-a974-b285e03c0c55","Type":"ContainerStarted","Data":"9b4843e4b148cfbebe1ac315d6231b9c9dee183c546e3c9adaf0cd06cd4cdcaa"} Oct 07 22:31:31 crc kubenswrapper[4871]: I1007 22:31:31.827839 4871 generic.go:334] "Generic (PLEG): container finished" podID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerID="70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc" exitCode=0 Oct 07 22:31:31 crc kubenswrapper[4871]: I1007 22:31:31.828080 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hl55" event={"ID":"874af2ed-f68b-458a-8a82-f3c433140b9c","Type":"ContainerDied","Data":"70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc"} Oct 07 22:31:31 crc kubenswrapper[4871]: I1007 22:31:31.828150 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hl55" event={"ID":"874af2ed-f68b-458a-8a82-f3c433140b9c","Type":"ContainerStarted","Data":"5c7e9b4bea4482a8096c70d1e64ddd28d86cd7427a6554892593b203724d51f8"} Oct 07 22:31:31 crc kubenswrapper[4871]: I1007 22:31:31.854028 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.853999024 podStartE2EDuration="2.853999024s" podCreationTimestamp="2025-10-07 22:31:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:31.85004886 +0000 UTC m=+1365.652746953" watchObservedRunningTime="2025-10-07 22:31:31.853999024 +0000 UTC m=+1365.656697107" Oct 07 22:31:32 crc kubenswrapper[4871]: I1007 22:31:32.843890 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerStarted","Data":"dbc726dfa2b8260803953e1a2042a77c88e3dd14b6404f0cc67de470fb14c6be"} Oct 07 22:31:32 crc kubenswrapper[4871]: I1007 22:31:32.844996 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 22:31:32 crc kubenswrapper[4871]: I1007 22:31:32.848840 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hl55" event={"ID":"874af2ed-f68b-458a-8a82-f3c433140b9c","Type":"ContainerStarted","Data":"68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf"} Oct 07 22:31:32 crc kubenswrapper[4871]: I1007 22:31:32.873318 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.291467575 podStartE2EDuration="6.873295866s" podCreationTimestamp="2025-10-07 22:31:26 +0000 UTC" firstStartedPulling="2025-10-07 22:31:27.832269868 +0000 UTC m=+1361.634967931" lastFinishedPulling="2025-10-07 22:31:32.414098139 +0000 UTC m=+1366.216796222" observedRunningTime="2025-10-07 22:31:32.871182201 +0000 UTC m=+1366.673880274" watchObservedRunningTime="2025-10-07 22:31:32.873295866 +0000 UTC m=+1366.675993969" Oct 07 22:31:33 crc kubenswrapper[4871]: I1007 22:31:33.866547 4871 generic.go:334] "Generic (PLEG): container finished" podID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerID="68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf" exitCode=0 Oct 07 22:31:33 crc kubenswrapper[4871]: I1007 22:31:33.867397 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hl55" event={"ID":"874af2ed-f68b-458a-8a82-f3c433140b9c","Type":"ContainerDied","Data":"68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf"} Oct 07 22:31:33 crc kubenswrapper[4871]: I1007 22:31:33.975393 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 07 22:31:34 crc kubenswrapper[4871]: I1007 22:31:34.885116 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hl55" event={"ID":"874af2ed-f68b-458a-8a82-f3c433140b9c","Type":"ContainerStarted","Data":"75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4"} Oct 07 22:31:34 crc kubenswrapper[4871]: I1007 22:31:34.914312 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4hl55" podStartSLOduration=2.094351766 podStartE2EDuration="4.914295203s" podCreationTimestamp="2025-10-07 22:31:30 +0000 UTC" firstStartedPulling="2025-10-07 22:31:31.833293919 +0000 UTC m=+1365.635992002" lastFinishedPulling="2025-10-07 22:31:34.653237366 +0000 UTC m=+1368.455935439" observedRunningTime="2025-10-07 22:31:34.90695859 +0000 UTC m=+1368.709656663" watchObservedRunningTime="2025-10-07 22:31:34.914295203 +0000 UTC m=+1368.716993276" Oct 07 22:31:35 crc kubenswrapper[4871]: I1007 22:31:35.152342 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 22:31:35 crc kubenswrapper[4871]: I1007 22:31:35.358791 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 22:31:35 crc kubenswrapper[4871]: I1007 22:31:35.360340 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 22:31:36 crc kubenswrapper[4871]: I1007 22:31:36.368038 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 22:31:36 crc kubenswrapper[4871]: I1007 22:31:36.368055 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 22:31:39 crc kubenswrapper[4871]: I1007 22:31:39.183540 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 22:31:39 crc kubenswrapper[4871]: I1007 22:31:39.184104 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 22:31:40 crc kubenswrapper[4871]: I1007 22:31:40.152239 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 22:31:40 crc kubenswrapper[4871]: I1007 22:31:40.205669 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 22:31:40 crc kubenswrapper[4871]: I1007 22:31:40.266163 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 22:31:40 crc kubenswrapper[4871]: I1007 22:31:40.267542 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 22:31:40 crc kubenswrapper[4871]: I1007 22:31:40.780261 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:40 crc kubenswrapper[4871]: I1007 22:31:40.780382 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:40 crc kubenswrapper[4871]: I1007 22:31:40.874096 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:41 crc kubenswrapper[4871]: I1007 22:31:41.048042 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 22:31:41 crc kubenswrapper[4871]: I1007 22:31:41.075330 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:41 crc kubenswrapper[4871]: I1007 22:31:41.138696 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hl55"] Oct 07 22:31:42 crc kubenswrapper[4871]: I1007 22:31:42.987256 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4hl55" podUID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerName="registry-server" containerID="cri-o://75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4" gracePeriod=2 Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.651341 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.805304 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98f72\" (UniqueName: \"kubernetes.io/projected/874af2ed-f68b-458a-8a82-f3c433140b9c-kube-api-access-98f72\") pod \"874af2ed-f68b-458a-8a82-f3c433140b9c\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.805392 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-catalog-content\") pod \"874af2ed-f68b-458a-8a82-f3c433140b9c\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.805481 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-utilities\") pod \"874af2ed-f68b-458a-8a82-f3c433140b9c\" (UID: \"874af2ed-f68b-458a-8a82-f3c433140b9c\") " Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.806958 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-utilities" (OuterVolumeSpecName: "utilities") pod "874af2ed-f68b-458a-8a82-f3c433140b9c" (UID: "874af2ed-f68b-458a-8a82-f3c433140b9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.821010 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/874af2ed-f68b-458a-8a82-f3c433140b9c-kube-api-access-98f72" (OuterVolumeSpecName: "kube-api-access-98f72") pod "874af2ed-f68b-458a-8a82-f3c433140b9c" (UID: "874af2ed-f68b-458a-8a82-f3c433140b9c"). InnerVolumeSpecName "kube-api-access-98f72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.822697 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "874af2ed-f68b-458a-8a82-f3c433140b9c" (UID: "874af2ed-f68b-458a-8a82-f3c433140b9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.909450 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98f72\" (UniqueName: \"kubernetes.io/projected/874af2ed-f68b-458a-8a82-f3c433140b9c-kube-api-access-98f72\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.909520 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:43 crc kubenswrapper[4871]: I1007 22:31:43.909537 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874af2ed-f68b-458a-8a82-f3c433140b9c-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.001265 4871 generic.go:334] "Generic (PLEG): container finished" podID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerID="75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4" exitCode=0 Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.001317 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hl55" event={"ID":"874af2ed-f68b-458a-8a82-f3c433140b9c","Type":"ContainerDied","Data":"75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4"} Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.001353 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hl55" event={"ID":"874af2ed-f68b-458a-8a82-f3c433140b9c","Type":"ContainerDied","Data":"5c7e9b4bea4482a8096c70d1e64ddd28d86cd7427a6554892593b203724d51f8"} Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.001374 4871 scope.go:117] "RemoveContainer" containerID="75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.001539 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hl55" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.046525 4871 scope.go:117] "RemoveContainer" containerID="68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.059186 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hl55"] Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.073010 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hl55"] Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.085902 4871 scope.go:117] "RemoveContainer" containerID="70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.148814 4871 scope.go:117] "RemoveContainer" containerID="75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4" Oct 07 22:31:44 crc kubenswrapper[4871]: E1007 22:31:44.149370 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4\": container with ID starting with 75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4 not found: ID does not exist" containerID="75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.149412 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4"} err="failed to get container status \"75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4\": rpc error: code = NotFound desc = could not find container \"75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4\": container with ID starting with 75f530f642731af51010e84f6884521ee027d51ebce714c5522ec9b571211fe4 not found: ID does not exist" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.149440 4871 scope.go:117] "RemoveContainer" containerID="68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf" Oct 07 22:31:44 crc kubenswrapper[4871]: E1007 22:31:44.150493 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf\": container with ID starting with 68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf not found: ID does not exist" containerID="68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.150524 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf"} err="failed to get container status \"68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf\": rpc error: code = NotFound desc = could not find container \"68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf\": container with ID starting with 68038d9edc5970a4bf7f7e0873c6d1c34256cd470b21b06d34ed88a8f2905adf not found: ID does not exist" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.150540 4871 scope.go:117] "RemoveContainer" containerID="70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc" Oct 07 22:31:44 crc kubenswrapper[4871]: E1007 22:31:44.151183 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc\": container with ID starting with 70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc not found: ID does not exist" containerID="70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc" Oct 07 22:31:44 crc kubenswrapper[4871]: I1007 22:31:44.151265 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc"} err="failed to get container status \"70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc\": rpc error: code = NotFound desc = could not find container \"70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc\": container with ID starting with 70589fd836f5949f3013d5069a9dfbb995f29e1eb25b788f413875e2a02d7dcc not found: ID does not exist" Oct 07 22:31:45 crc kubenswrapper[4871]: I1007 22:31:45.003386 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="874af2ed-f68b-458a-8a82-f3c433140b9c" path="/var/lib/kubelet/pods/874af2ed-f68b-458a-8a82-f3c433140b9c/volumes" Oct 07 22:31:45 crc kubenswrapper[4871]: I1007 22:31:45.367421 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 22:31:45 crc kubenswrapper[4871]: I1007 22:31:45.368054 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 22:31:45 crc kubenswrapper[4871]: I1007 22:31:45.374498 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 22:31:45 crc kubenswrapper[4871]: I1007 22:31:45.375269 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 22:31:46 crc kubenswrapper[4871]: I1007 22:31:46.821542 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:46 crc kubenswrapper[4871]: I1007 22:31:46.989580 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9d65\" (UniqueName: \"kubernetes.io/projected/74fab45c-4927-4ef6-b85e-05158f2af03e-kube-api-access-n9d65\") pod \"74fab45c-4927-4ef6-b85e-05158f2af03e\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " Oct 07 22:31:46 crc kubenswrapper[4871]: I1007 22:31:46.989831 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-combined-ca-bundle\") pod \"74fab45c-4927-4ef6-b85e-05158f2af03e\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " Oct 07 22:31:46 crc kubenswrapper[4871]: I1007 22:31:46.990164 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-config-data\") pod \"74fab45c-4927-4ef6-b85e-05158f2af03e\" (UID: \"74fab45c-4927-4ef6-b85e-05158f2af03e\") " Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.015282 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74fab45c-4927-4ef6-b85e-05158f2af03e-kube-api-access-n9d65" (OuterVolumeSpecName: "kube-api-access-n9d65") pod "74fab45c-4927-4ef6-b85e-05158f2af03e" (UID: "74fab45c-4927-4ef6-b85e-05158f2af03e"). InnerVolumeSpecName "kube-api-access-n9d65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.045290 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74fab45c-4927-4ef6-b85e-05158f2af03e" (UID: "74fab45c-4927-4ef6-b85e-05158f2af03e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.045374 4871 generic.go:334] "Generic (PLEG): container finished" podID="74fab45c-4927-4ef6-b85e-05158f2af03e" containerID="e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49" exitCode=137 Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.045482 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.045916 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-config-data" (OuterVolumeSpecName: "config-data") pod "74fab45c-4927-4ef6-b85e-05158f2af03e" (UID: "74fab45c-4927-4ef6-b85e-05158f2af03e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.093703 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.094207 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9d65\" (UniqueName: \"kubernetes.io/projected/74fab45c-4927-4ef6-b85e-05158f2af03e-kube-api-access-n9d65\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.094218 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fab45c-4927-4ef6-b85e-05158f2af03e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.098872 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"74fab45c-4927-4ef6-b85e-05158f2af03e","Type":"ContainerDied","Data":"e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49"} Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.098913 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"74fab45c-4927-4ef6-b85e-05158f2af03e","Type":"ContainerDied","Data":"b7af9a17ddcc74916d39d4bbc91d614b59c3b9452997c1a663aad4cb5590aa06"} Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.098933 4871 scope.go:117] "RemoveContainer" containerID="e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.144238 4871 scope.go:117] "RemoveContainer" containerID="e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49" Oct 07 22:31:47 crc kubenswrapper[4871]: E1007 22:31:47.144918 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49\": container with ID starting with e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49 not found: ID does not exist" containerID="e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.144959 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49"} err="failed to get container status \"e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49\": rpc error: code = NotFound desc = could not find container \"e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49\": container with ID starting with e311c9052f80cc487192861dee0d8f5e423edda399127efac7d647b14eefce49 not found: ID does not exist" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.409563 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.430656 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.444596 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:31:47 crc kubenswrapper[4871]: E1007 22:31:47.445276 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerName="extract-utilities" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.445313 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerName="extract-utilities" Oct 07 22:31:47 crc kubenswrapper[4871]: E1007 22:31:47.445340 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerName="extract-content" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.445354 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerName="extract-content" Oct 07 22:31:47 crc kubenswrapper[4871]: E1007 22:31:47.445404 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerName="registry-server" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.445418 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerName="registry-server" Oct 07 22:31:47 crc kubenswrapper[4871]: E1007 22:31:47.445460 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74fab45c-4927-4ef6-b85e-05158f2af03e" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.445473 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="74fab45c-4927-4ef6-b85e-05158f2af03e" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.445836 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="874af2ed-f68b-458a-8a82-f3c433140b9c" containerName="registry-server" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.445884 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="74fab45c-4927-4ef6-b85e-05158f2af03e" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.447637 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.450719 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.451057 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.451085 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.460909 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.501676 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.501940 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.501976 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.502004 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fmm8\" (UniqueName: \"kubernetes.io/projected/c9e4de98-7d7d-43af-a76b-abf1774c620d-kube-api-access-6fmm8\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.502092 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.604026 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.604435 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.604483 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.604547 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fmm8\" (UniqueName: \"kubernetes.io/projected/c9e4de98-7d7d-43af-a76b-abf1774c620d-kube-api-access-6fmm8\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.604625 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.610620 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.610919 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.620420 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.624523 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.628222 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fmm8\" (UniqueName: \"kubernetes.io/projected/c9e4de98-7d7d-43af-a76b-abf1774c620d-kube-api-access-6fmm8\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:47 crc kubenswrapper[4871]: I1007 22:31:47.793113 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:48 crc kubenswrapper[4871]: I1007 22:31:48.271912 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:31:48 crc kubenswrapper[4871]: W1007 22:31:48.275526 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9e4de98_7d7d_43af_a76b_abf1774c620d.slice/crio-f3956bb5766dd205fa72ec949c6e4fca7a5b0c2ca288def1748431f4ffa0258d WatchSource:0}: Error finding container f3956bb5766dd205fa72ec949c6e4fca7a5b0c2ca288def1748431f4ffa0258d: Status 404 returned error can't find the container with id f3956bb5766dd205fa72ec949c6e4fca7a5b0c2ca288def1748431f4ffa0258d Oct 07 22:31:48 crc kubenswrapper[4871]: I1007 22:31:48.996634 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74fab45c-4927-4ef6-b85e-05158f2af03e" path="/var/lib/kubelet/pods/74fab45c-4927-4ef6-b85e-05158f2af03e/volumes" Oct 07 22:31:49 crc kubenswrapper[4871]: I1007 22:31:49.082027 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c9e4de98-7d7d-43af-a76b-abf1774c620d","Type":"ContainerStarted","Data":"2ff1e6c6e2b4bf5a4e4fe91286071d2d9c47cf4d7640e8db226f47e4bb0157a4"} Oct 07 22:31:49 crc kubenswrapper[4871]: I1007 22:31:49.082094 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c9e4de98-7d7d-43af-a76b-abf1774c620d","Type":"ContainerStarted","Data":"f3956bb5766dd205fa72ec949c6e4fca7a5b0c2ca288def1748431f4ffa0258d"} Oct 07 22:31:49 crc kubenswrapper[4871]: I1007 22:31:49.106823 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.106772661 podStartE2EDuration="2.106772661s" podCreationTimestamp="2025-10-07 22:31:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:49.099779716 +0000 UTC m=+1382.902477789" watchObservedRunningTime="2025-10-07 22:31:49.106772661 +0000 UTC m=+1382.909470754" Oct 07 22:31:49 crc kubenswrapper[4871]: I1007 22:31:49.188188 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 22:31:49 crc kubenswrapper[4871]: I1007 22:31:49.189906 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 22:31:49 crc kubenswrapper[4871]: I1007 22:31:49.191366 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 22:31:49 crc kubenswrapper[4871]: I1007 22:31:49.206471 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.093898 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.100593 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.367486 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-phtsg"] Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.371440 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.389686 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-phtsg"] Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.480400 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-sb\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.480471 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-svc\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.480520 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-swift-storage-0\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.480551 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-nb\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.480622 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd57k\" (UniqueName: \"kubernetes.io/projected/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-kube-api-access-hd57k\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.480701 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-config\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.583431 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-svc\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.584777 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-sb\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.585440 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-swift-storage-0\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.586060 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-nb\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.585366 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-sb\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.585997 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-swift-storage-0\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.584718 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-svc\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.586229 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd57k\" (UniqueName: \"kubernetes.io/projected/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-kube-api-access-hd57k\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.586654 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-config\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.586884 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-nb\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.587373 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-config\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.613269 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd57k\" (UniqueName: \"kubernetes.io/projected/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-kube-api-access-hd57k\") pod \"dnsmasq-dns-5967cc9597-phtsg\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:50 crc kubenswrapper[4871]: I1007 22:31:50.702588 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:51 crc kubenswrapper[4871]: I1007 22:31:51.219775 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-phtsg"] Oct 07 22:31:51 crc kubenswrapper[4871]: W1007 22:31:51.231264 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5d55d4b_6e19_46ed_9ca8_04ddab7c8b2e.slice/crio-b8b4f077f6ba2b4943c091859f32d76cb65df0d33c978f83c2370116ac851bdb WatchSource:0}: Error finding container b8b4f077f6ba2b4943c091859f32d76cb65df0d33c978f83c2370116ac851bdb: Status 404 returned error can't find the container with id b8b4f077f6ba2b4943c091859f32d76cb65df0d33c978f83c2370116ac851bdb Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.115151 4871 generic.go:334] "Generic (PLEG): container finished" podID="e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" containerID="33f7ea0548fe776e435826efc31be59ac4fe01983421306e3a6bb36a95b9f477" exitCode=0 Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.115452 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" event={"ID":"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e","Type":"ContainerDied","Data":"33f7ea0548fe776e435826efc31be59ac4fe01983421306e3a6bb36a95b9f477"} Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.115510 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" event={"ID":"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e","Type":"ContainerStarted","Data":"b8b4f077f6ba2b4943c091859f32d76cb65df0d33c978f83c2370116ac851bdb"} Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.151460 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.151849 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="ceilometer-central-agent" containerID="cri-o://83982a92e0f72b913a826e96cf85b93a00c47ee62f56aeb66b806d9050b085f4" gracePeriod=30 Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.152530 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="proxy-httpd" containerID="cri-o://dbc726dfa2b8260803953e1a2042a77c88e3dd14b6404f0cc67de470fb14c6be" gracePeriod=30 Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.152620 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="sg-core" containerID="cri-o://1ddd5996c55ae6aba76392b6bb32551ae5d81fd8bdb6f3a5fc16898023ab622c" gracePeriod=30 Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.152539 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="ceilometer-notification-agent" containerID="cri-o://60d46800f21956a308b11718acac3ac7f483202813d32b1ad29888a3936484ea" gracePeriod=30 Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.168747 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.193:3000/\": EOF" Oct 07 22:31:52 crc kubenswrapper[4871]: I1007 22:31:52.793426 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.027006 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.128823 4871 generic.go:334] "Generic (PLEG): container finished" podID="237934fb-75da-40fe-a475-bcc78f45daa3" containerID="dbc726dfa2b8260803953e1a2042a77c88e3dd14b6404f0cc67de470fb14c6be" exitCode=0 Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.128867 4871 generic.go:334] "Generic (PLEG): container finished" podID="237934fb-75da-40fe-a475-bcc78f45daa3" containerID="1ddd5996c55ae6aba76392b6bb32551ae5d81fd8bdb6f3a5fc16898023ab622c" exitCode=2 Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.128876 4871 generic.go:334] "Generic (PLEG): container finished" podID="237934fb-75da-40fe-a475-bcc78f45daa3" containerID="83982a92e0f72b913a826e96cf85b93a00c47ee62f56aeb66b806d9050b085f4" exitCode=0 Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.128895 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerDied","Data":"dbc726dfa2b8260803953e1a2042a77c88e3dd14b6404f0cc67de470fb14c6be"} Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.128982 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerDied","Data":"1ddd5996c55ae6aba76392b6bb32551ae5d81fd8bdb6f3a5fc16898023ab622c"} Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.129002 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerDied","Data":"83982a92e0f72b913a826e96cf85b93a00c47ee62f56aeb66b806d9050b085f4"} Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.131557 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" event={"ID":"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e","Type":"ContainerStarted","Data":"36507cab7159abb01d536c353f20262fe1f5bacd4aef76bb370e449a8777f532"} Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.131724 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.131905 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-log" containerID="cri-o://f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa" gracePeriod=30 Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.132001 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-api" containerID="cri-o://b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee" gracePeriod=30 Oct 07 22:31:53 crc kubenswrapper[4871]: I1007 22:31:53.160319 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" podStartSLOduration=3.160299884 podStartE2EDuration="3.160299884s" podCreationTimestamp="2025-10-07 22:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:53.155688313 +0000 UTC m=+1386.958386406" watchObservedRunningTime="2025-10-07 22:31:53.160299884 +0000 UTC m=+1386.962997957" Oct 07 22:31:54 crc kubenswrapper[4871]: I1007 22:31:54.145980 4871 generic.go:334] "Generic (PLEG): container finished" podID="1812a7df-9924-454b-91a7-28df2d766ab9" containerID="f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa" exitCode=143 Oct 07 22:31:54 crc kubenswrapper[4871]: I1007 22:31:54.146068 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1812a7df-9924-454b-91a7-28df2d766ab9","Type":"ContainerDied","Data":"f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa"} Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.185007 4871 generic.go:334] "Generic (PLEG): container finished" podID="237934fb-75da-40fe-a475-bcc78f45daa3" containerID="60d46800f21956a308b11718acac3ac7f483202813d32b1ad29888a3936484ea" exitCode=0 Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.185137 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerDied","Data":"60d46800f21956a308b11718acac3ac7f483202813d32b1ad29888a3936484ea"} Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.497508 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.661007 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-ceilometer-tls-certs\") pod \"237934fb-75da-40fe-a475-bcc78f45daa3\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.661140 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x9dw\" (UniqueName: \"kubernetes.io/projected/237934fb-75da-40fe-a475-bcc78f45daa3-kube-api-access-7x9dw\") pod \"237934fb-75da-40fe-a475-bcc78f45daa3\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.661337 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-combined-ca-bundle\") pod \"237934fb-75da-40fe-a475-bcc78f45daa3\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.661426 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-sg-core-conf-yaml\") pod \"237934fb-75da-40fe-a475-bcc78f45daa3\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.661494 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-config-data\") pod \"237934fb-75da-40fe-a475-bcc78f45daa3\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.661579 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-run-httpd\") pod \"237934fb-75da-40fe-a475-bcc78f45daa3\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.661635 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-log-httpd\") pod \"237934fb-75da-40fe-a475-bcc78f45daa3\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.661671 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-scripts\") pod \"237934fb-75da-40fe-a475-bcc78f45daa3\" (UID: \"237934fb-75da-40fe-a475-bcc78f45daa3\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.664062 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "237934fb-75da-40fe-a475-bcc78f45daa3" (UID: "237934fb-75da-40fe-a475-bcc78f45daa3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.664281 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "237934fb-75da-40fe-a475-bcc78f45daa3" (UID: "237934fb-75da-40fe-a475-bcc78f45daa3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.666806 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.666836 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/237934fb-75da-40fe-a475-bcc78f45daa3-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.671546 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-scripts" (OuterVolumeSpecName: "scripts") pod "237934fb-75da-40fe-a475-bcc78f45daa3" (UID: "237934fb-75da-40fe-a475-bcc78f45daa3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.674513 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/237934fb-75da-40fe-a475-bcc78f45daa3-kube-api-access-7x9dw" (OuterVolumeSpecName: "kube-api-access-7x9dw") pod "237934fb-75da-40fe-a475-bcc78f45daa3" (UID: "237934fb-75da-40fe-a475-bcc78f45daa3"). InnerVolumeSpecName "kube-api-access-7x9dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.724194 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "237934fb-75da-40fe-a475-bcc78f45daa3" (UID: "237934fb-75da-40fe-a475-bcc78f45daa3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.726692 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "237934fb-75da-40fe-a475-bcc78f45daa3" (UID: "237934fb-75da-40fe-a475-bcc78f45daa3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.755987 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "237934fb-75da-40fe-a475-bcc78f45daa3" (UID: "237934fb-75da-40fe-a475-bcc78f45daa3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.768399 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.768435 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.768445 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.768456 4871 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.768465 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x9dw\" (UniqueName: \"kubernetes.io/projected/237934fb-75da-40fe-a475-bcc78f45daa3-kube-api-access-7x9dw\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.800038 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.810720 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-config-data" (OuterVolumeSpecName: "config-data") pod "237934fb-75da-40fe-a475-bcc78f45daa3" (UID: "237934fb-75da-40fe-a475-bcc78f45daa3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.868902 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx462\" (UniqueName: \"kubernetes.io/projected/1812a7df-9924-454b-91a7-28df2d766ab9-kube-api-access-tx462\") pod \"1812a7df-9924-454b-91a7-28df2d766ab9\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.869248 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-combined-ca-bundle\") pod \"1812a7df-9924-454b-91a7-28df2d766ab9\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.869481 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1812a7df-9924-454b-91a7-28df2d766ab9-logs\") pod \"1812a7df-9924-454b-91a7-28df2d766ab9\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.869555 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-config-data\") pod \"1812a7df-9924-454b-91a7-28df2d766ab9\" (UID: \"1812a7df-9924-454b-91a7-28df2d766ab9\") " Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.869976 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1812a7df-9924-454b-91a7-28df2d766ab9-logs" (OuterVolumeSpecName: "logs") pod "1812a7df-9924-454b-91a7-28df2d766ab9" (UID: "1812a7df-9924-454b-91a7-28df2d766ab9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.872021 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1812a7df-9924-454b-91a7-28df2d766ab9-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.872093 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237934fb-75da-40fe-a475-bcc78f45daa3-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.879004 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1812a7df-9924-454b-91a7-28df2d766ab9-kube-api-access-tx462" (OuterVolumeSpecName: "kube-api-access-tx462") pod "1812a7df-9924-454b-91a7-28df2d766ab9" (UID: "1812a7df-9924-454b-91a7-28df2d766ab9"). InnerVolumeSpecName "kube-api-access-tx462". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.897119 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-config-data" (OuterVolumeSpecName: "config-data") pod "1812a7df-9924-454b-91a7-28df2d766ab9" (UID: "1812a7df-9924-454b-91a7-28df2d766ab9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.899298 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1812a7df-9924-454b-91a7-28df2d766ab9" (UID: "1812a7df-9924-454b-91a7-28df2d766ab9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.974462 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx462\" (UniqueName: \"kubernetes.io/projected/1812a7df-9924-454b-91a7-28df2d766ab9-kube-api-access-tx462\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.974641 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:56 crc kubenswrapper[4871]: I1007 22:31:56.974770 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1812a7df-9924-454b-91a7-28df2d766ab9-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.202527 4871 generic.go:334] "Generic (PLEG): container finished" podID="1812a7df-9924-454b-91a7-28df2d766ab9" containerID="b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee" exitCode=0 Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.202676 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1812a7df-9924-454b-91a7-28df2d766ab9","Type":"ContainerDied","Data":"b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee"} Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.202737 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1812a7df-9924-454b-91a7-28df2d766ab9","Type":"ContainerDied","Data":"9859c1b2fc6a984dab270551656e10fb27203831664946bdb9e32dc037e726ba"} Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.202777 4871 scope.go:117] "RemoveContainer" containerID="b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.204430 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.210860 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"237934fb-75da-40fe-a475-bcc78f45daa3","Type":"ContainerDied","Data":"d429507f75776fc4264a5ec8876777756d9fb21aa92148013829ca0100ac6147"} Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.211017 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.291856 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.293640 4871 scope.go:117] "RemoveContainer" containerID="f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.322876 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.379950 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.387843 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:57 crc kubenswrapper[4871]: E1007 22:31:57.388585 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-api" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.395039 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-api" Oct 07 22:31:57 crc kubenswrapper[4871]: E1007 22:31:57.395221 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="sg-core" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.395277 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="sg-core" Oct 07 22:31:57 crc kubenswrapper[4871]: E1007 22:31:57.395334 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="proxy-httpd" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.395384 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="proxy-httpd" Oct 07 22:31:57 crc kubenswrapper[4871]: E1007 22:31:57.395452 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="ceilometer-central-agent" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.395509 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="ceilometer-central-agent" Oct 07 22:31:57 crc kubenswrapper[4871]: E1007 22:31:57.395575 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-log" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.395639 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-log" Oct 07 22:31:57 crc kubenswrapper[4871]: E1007 22:31:57.395707 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="ceilometer-notification-agent" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.395759 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="ceilometer-notification-agent" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.396141 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="sg-core" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.396212 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="ceilometer-notification-agent" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.396272 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="proxy-httpd" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.396335 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-api" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.396400 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" containerName="ceilometer-central-agent" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.396465 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" containerName="nova-api-log" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.397679 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.406372 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.407048 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.411069 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.416318 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.425397 4871 scope.go:117] "RemoveContainer" containerID="b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee" Oct 07 22:31:57 crc kubenswrapper[4871]: E1007 22:31:57.430249 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee\": container with ID starting with b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee not found: ID does not exist" containerID="b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.430375 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee"} err="failed to get container status \"b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee\": rpc error: code = NotFound desc = could not find container \"b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee\": container with ID starting with b30e27f0647d2275470138a60de3360f52abb2df71230f1ba460bbdbaf44d5ee not found: ID does not exist" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.430500 4871 scope.go:117] "RemoveContainer" containerID="f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.432966 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:57 crc kubenswrapper[4871]: E1007 22:31:57.436390 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa\": container with ID starting with f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa not found: ID does not exist" containerID="f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.436479 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa"} err="failed to get container status \"f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa\": rpc error: code = NotFound desc = could not find container \"f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa\": container with ID starting with f14d3d200821aca63b2810449030214bc4e8f4a625e67e255229723e080bb3fa not found: ID does not exist" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.436645 4871 scope.go:117] "RemoveContainer" containerID="dbc726dfa2b8260803953e1a2042a77c88e3dd14b6404f0cc67de470fb14c6be" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.442874 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.445527 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.450246 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.453360 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.454171 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.454244 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491584 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491662 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491682 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-scripts\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491716 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ptvb\" (UniqueName: \"kubernetes.io/projected/e71edefc-c645-4e74-8077-42df9e751ce0-kube-api-access-6ptvb\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491762 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491784 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-public-tls-certs\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491833 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-log-httpd\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491873 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491899 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxhtl\" (UniqueName: \"kubernetes.io/projected/03a6b072-112e-4e9d-8f58-669d4127c21b-kube-api-access-mxhtl\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491927 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-run-httpd\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491944 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-config-data\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.491991 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.492010 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03a6b072-112e-4e9d-8f58-669d4127c21b-logs\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.492170 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-config-data\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.504922 4871 scope.go:117] "RemoveContainer" containerID="1ddd5996c55ae6aba76392b6bb32551ae5d81fd8bdb6f3a5fc16898023ab622c" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.524424 4871 scope.go:117] "RemoveContainer" containerID="60d46800f21956a308b11718acac3ac7f483202813d32b1ad29888a3936484ea" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.545332 4871 scope.go:117] "RemoveContainer" containerID="83982a92e0f72b913a826e96cf85b93a00c47ee62f56aeb66b806d9050b085f4" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593651 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593699 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxhtl\" (UniqueName: \"kubernetes.io/projected/03a6b072-112e-4e9d-8f58-669d4127c21b-kube-api-access-mxhtl\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593733 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-run-httpd\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593752 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-config-data\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593804 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593820 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03a6b072-112e-4e9d-8f58-669d4127c21b-logs\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593846 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-config-data\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593881 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593910 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593927 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-scripts\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.593946 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ptvb\" (UniqueName: \"kubernetes.io/projected/e71edefc-c645-4e74-8077-42df9e751ce0-kube-api-access-6ptvb\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.594032 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.594049 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-public-tls-certs\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.594065 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-log-httpd\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.594573 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-log-httpd\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.594978 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-run-httpd\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.597729 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03a6b072-112e-4e9d-8f58-669d4127c21b-logs\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.600319 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.601122 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.602681 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.603865 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.604049 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-scripts\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.604833 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-config-data\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.605552 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-public-tls-certs\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.606387 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-config-data\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.612913 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxhtl\" (UniqueName: \"kubernetes.io/projected/03a6b072-112e-4e9d-8f58-669d4127c21b-kube-api-access-mxhtl\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.615169 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.630579 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ptvb\" (UniqueName: \"kubernetes.io/projected/e71edefc-c645-4e74-8077-42df9e751ce0-kube-api-access-6ptvb\") pod \"ceilometer-0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.746012 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.777908 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.793296 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:57 crc kubenswrapper[4871]: I1007 22:31:57.820386 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.248832 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.338727 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.345737 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.384192 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.502472 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-rnkr2"] Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.504506 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.508106 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.508322 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.512836 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rnkr2"] Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.528179 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.528340 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-scripts\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.528410 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htvr9\" (UniqueName: \"kubernetes.io/projected/7626d31e-6b79-44a4-b29d-0912f08ccbc7-kube-api-access-htvr9\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.528513 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-config-data\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.630730 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.630806 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-scripts\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.630839 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htvr9\" (UniqueName: \"kubernetes.io/projected/7626d31e-6b79-44a4-b29d-0912f08ccbc7-kube-api-access-htvr9\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.630877 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-config-data\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.637887 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-config-data\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.637930 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-scripts\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.638301 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.650935 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htvr9\" (UniqueName: \"kubernetes.io/projected/7626d31e-6b79-44a4-b29d-0912f08ccbc7-kube-api-access-htvr9\") pod \"nova-cell1-cell-mapping-rnkr2\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.873332 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.997279 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1812a7df-9924-454b-91a7-28df2d766ab9" path="/var/lib/kubelet/pods/1812a7df-9924-454b-91a7-28df2d766ab9/volumes" Oct 07 22:31:58 crc kubenswrapper[4871]: I1007 22:31:58.997942 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="237934fb-75da-40fe-a475-bcc78f45daa3" path="/var/lib/kubelet/pods/237934fb-75da-40fe-a475-bcc78f45daa3/volumes" Oct 07 22:31:59 crc kubenswrapper[4871]: I1007 22:31:59.250754 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03a6b072-112e-4e9d-8f58-669d4127c21b","Type":"ContainerStarted","Data":"267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d"} Oct 07 22:31:59 crc kubenswrapper[4871]: I1007 22:31:59.251317 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03a6b072-112e-4e9d-8f58-669d4127c21b","Type":"ContainerStarted","Data":"5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec"} Oct 07 22:31:59 crc kubenswrapper[4871]: I1007 22:31:59.251342 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03a6b072-112e-4e9d-8f58-669d4127c21b","Type":"ContainerStarted","Data":"059eb4ebcd4dcc439fcde0c02c2a64e37ec1c4fae9d1c51a9d477f7a6be80bb5"} Oct 07 22:31:59 crc kubenswrapper[4871]: I1007 22:31:59.257999 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerStarted","Data":"33c6f7c23038063d0da214f2c17dddbf223e200146c8e1e89dee21bab58b13a5"} Oct 07 22:31:59 crc kubenswrapper[4871]: I1007 22:31:59.278343 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.278311724 podStartE2EDuration="2.278311724s" podCreationTimestamp="2025-10-07 22:31:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:31:59.275004457 +0000 UTC m=+1393.077702570" watchObservedRunningTime="2025-10-07 22:31:59.278311724 +0000 UTC m=+1393.081009817" Oct 07 22:31:59 crc kubenswrapper[4871]: I1007 22:31:59.347027 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rnkr2"] Oct 07 22:32:00 crc kubenswrapper[4871]: I1007 22:32:00.271935 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerStarted","Data":"b2140ff400665da608cd7a343da108840518d021eb0a870d34ed4a85d295f742"} Oct 07 22:32:00 crc kubenswrapper[4871]: I1007 22:32:00.274346 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rnkr2" event={"ID":"7626d31e-6b79-44a4-b29d-0912f08ccbc7","Type":"ContainerStarted","Data":"41dcab2800cab633133f6b01f760c087220835afec33308c322b94a96c4c33d9"} Oct 07 22:32:00 crc kubenswrapper[4871]: I1007 22:32:00.274395 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rnkr2" event={"ID":"7626d31e-6b79-44a4-b29d-0912f08ccbc7","Type":"ContainerStarted","Data":"5a5dc9e77581df8709360f12e5b2586e64c62f0fa2a22c08105a04c0d1abd533"} Oct 07 22:32:00 crc kubenswrapper[4871]: I1007 22:32:00.295282 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-rnkr2" podStartSLOduration=2.295257823 podStartE2EDuration="2.295257823s" podCreationTimestamp="2025-10-07 22:31:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:32:00.290323333 +0000 UTC m=+1394.093021426" watchObservedRunningTime="2025-10-07 22:32:00.295257823 +0000 UTC m=+1394.097955906" Oct 07 22:32:00 crc kubenswrapper[4871]: I1007 22:32:00.703963 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:32:00 crc kubenswrapper[4871]: I1007 22:32:00.768037 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-z7nsb"] Oct 07 22:32:00 crc kubenswrapper[4871]: I1007 22:32:00.768368 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" podUID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" containerName="dnsmasq-dns" containerID="cri-o://ac7926263bebb46db052a48e0d59ffb573ae16e631980ef5e58f7119d8da1a49" gracePeriod=10 Oct 07 22:32:01 crc kubenswrapper[4871]: E1007 22:32:01.001749 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bbe7cd5_b5d5_49ea_9955_72c1d41f3e4a.slice/crio-ac7926263bebb46db052a48e0d59ffb573ae16e631980ef5e58f7119d8da1a49.scope\": RecentStats: unable to find data in memory cache]" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.315855 4871 generic.go:334] "Generic (PLEG): container finished" podID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" containerID="ac7926263bebb46db052a48e0d59ffb573ae16e631980ef5e58f7119d8da1a49" exitCode=0 Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.316343 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" event={"ID":"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a","Type":"ContainerDied","Data":"ac7926263bebb46db052a48e0d59ffb573ae16e631980ef5e58f7119d8da1a49"} Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.316387 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" event={"ID":"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a","Type":"ContainerDied","Data":"805dd6bbbc6336267286e4bb1f92a0839026a7bd2399b71b267ae58f9b95d410"} Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.316399 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="805dd6bbbc6336267286e4bb1f92a0839026a7bd2399b71b267ae58f9b95d410" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.341466 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerStarted","Data":"f02b2caa39b480dc045d7e7262686751f44a5f1546d461810390f4aa68b67b32"} Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.355682 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.399666 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-config\") pod \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.399755 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-swift-storage-0\") pod \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.399817 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-svc\") pod \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.399923 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-sb\") pod \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.399950 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-nb\") pod \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.400171 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zspj\" (UniqueName: \"kubernetes.io/projected/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-kube-api-access-7zspj\") pod \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\" (UID: \"2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a\") " Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.406013 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-kube-api-access-7zspj" (OuterVolumeSpecName: "kube-api-access-7zspj") pod "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" (UID: "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a"). InnerVolumeSpecName "kube-api-access-7zspj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.461680 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" (UID: "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.463124 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" (UID: "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.464962 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" (UID: "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.466371 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" (UID: "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.466404 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-config" (OuterVolumeSpecName: "config") pod "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" (UID: "2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.503215 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zspj\" (UniqueName: \"kubernetes.io/projected/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-kube-api-access-7zspj\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.503261 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.503272 4871 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.503281 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.503291 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:01 crc kubenswrapper[4871]: I1007 22:32:01.503301 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.353756 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerStarted","Data":"b1bc98f764cc4f92e14878e9fcdb1b85ceb45251562f19cc562511bd4c01c7c5"} Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.353773 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.388782 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-z7nsb"] Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.405572 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-z7nsb"] Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.773386 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hfjj4"] Oct 07 22:32:02 crc kubenswrapper[4871]: E1007 22:32:02.774001 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" containerName="dnsmasq-dns" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.774033 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" containerName="dnsmasq-dns" Oct 07 22:32:02 crc kubenswrapper[4871]: E1007 22:32:02.774063 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" containerName="init" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.774075 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" containerName="init" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.774436 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" containerName="dnsmasq-dns" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.776841 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.792858 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hfjj4"] Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.834918 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-utilities\") pod \"community-operators-hfjj4\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.835176 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t26v\" (UniqueName: \"kubernetes.io/projected/f82d7a4d-d310-4f1d-a382-757ee0b5b794-kube-api-access-6t26v\") pod \"community-operators-hfjj4\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.835466 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-catalog-content\") pod \"community-operators-hfjj4\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.936886 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-catalog-content\") pod \"community-operators-hfjj4\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.937001 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-utilities\") pod \"community-operators-hfjj4\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.937059 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t26v\" (UniqueName: \"kubernetes.io/projected/f82d7a4d-d310-4f1d-a382-757ee0b5b794-kube-api-access-6t26v\") pod \"community-operators-hfjj4\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.937356 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-catalog-content\") pod \"community-operators-hfjj4\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.937499 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-utilities\") pod \"community-operators-hfjj4\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.959231 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t26v\" (UniqueName: \"kubernetes.io/projected/f82d7a4d-d310-4f1d-a382-757ee0b5b794-kube-api-access-6t26v\") pod \"community-operators-hfjj4\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:02 crc kubenswrapper[4871]: I1007 22:32:02.994492 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" path="/var/lib/kubelet/pods/2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a/volumes" Oct 07 22:32:03 crc kubenswrapper[4871]: I1007 22:32:03.155023 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:03 crc kubenswrapper[4871]: I1007 22:32:03.380420 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerStarted","Data":"457a86497864d11e171b8cdea112cccb3d4066b22a76ceab718229ee7ea9a56c"} Oct 07 22:32:03 crc kubenswrapper[4871]: I1007 22:32:03.382379 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 22:32:03 crc kubenswrapper[4871]: I1007 22:32:03.416765 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.123490145 podStartE2EDuration="6.416723683s" podCreationTimestamp="2025-10-07 22:31:57 +0000 UTC" firstStartedPulling="2025-10-07 22:31:58.338486326 +0000 UTC m=+1392.141184399" lastFinishedPulling="2025-10-07 22:32:02.631719824 +0000 UTC m=+1396.434417937" observedRunningTime="2025-10-07 22:32:03.40748111 +0000 UTC m=+1397.210179183" watchObservedRunningTime="2025-10-07 22:32:03.416723683 +0000 UTC m=+1397.219421746" Oct 07 22:32:03 crc kubenswrapper[4871]: I1007 22:32:03.690252 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hfjj4"] Oct 07 22:32:04 crc kubenswrapper[4871]: I1007 22:32:04.398874 4871 generic.go:334] "Generic (PLEG): container finished" podID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerID="ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe" exitCode=0 Oct 07 22:32:04 crc kubenswrapper[4871]: I1007 22:32:04.399044 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfjj4" event={"ID":"f82d7a4d-d310-4f1d-a382-757ee0b5b794","Type":"ContainerDied","Data":"ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe"} Oct 07 22:32:04 crc kubenswrapper[4871]: I1007 22:32:04.399455 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfjj4" event={"ID":"f82d7a4d-d310-4f1d-a382-757ee0b5b794","Type":"ContainerStarted","Data":"c9eae584ffd66e6de840c57434c9533541750b6f7c06b5f61b0d6ddf36466984"} Oct 07 22:32:05 crc kubenswrapper[4871]: I1007 22:32:05.412980 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfjj4" event={"ID":"f82d7a4d-d310-4f1d-a382-757ee0b5b794","Type":"ContainerStarted","Data":"59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45"} Oct 07 22:32:05 crc kubenswrapper[4871]: I1007 22:32:05.418961 4871 generic.go:334] "Generic (PLEG): container finished" podID="7626d31e-6b79-44a4-b29d-0912f08ccbc7" containerID="41dcab2800cab633133f6b01f760c087220835afec33308c322b94a96c4c33d9" exitCode=0 Oct 07 22:32:05 crc kubenswrapper[4871]: I1007 22:32:05.419014 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rnkr2" event={"ID":"7626d31e-6b79-44a4-b29d-0912f08ccbc7","Type":"ContainerDied","Data":"41dcab2800cab633133f6b01f760c087220835afec33308c322b94a96c4c33d9"} Oct 07 22:32:06 crc kubenswrapper[4871]: I1007 22:32:06.127270 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-64d8d96789-z7nsb" podUID="2bbe7cd5-b5d5-49ea-9955-72c1d41f3e4a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.187:5353: i/o timeout" Oct 07 22:32:06 crc kubenswrapper[4871]: I1007 22:32:06.434951 4871 generic.go:334] "Generic (PLEG): container finished" podID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerID="59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45" exitCode=0 Oct 07 22:32:06 crc kubenswrapper[4871]: I1007 22:32:06.435058 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfjj4" event={"ID":"f82d7a4d-d310-4f1d-a382-757ee0b5b794","Type":"ContainerDied","Data":"59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45"} Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:06.999724 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.058901 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-combined-ca-bundle\") pod \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.059265 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-config-data\") pod \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.059497 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-scripts\") pod \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.059578 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htvr9\" (UniqueName: \"kubernetes.io/projected/7626d31e-6b79-44a4-b29d-0912f08ccbc7-kube-api-access-htvr9\") pod \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\" (UID: \"7626d31e-6b79-44a4-b29d-0912f08ccbc7\") " Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.067759 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-scripts" (OuterVolumeSpecName: "scripts") pod "7626d31e-6b79-44a4-b29d-0912f08ccbc7" (UID: "7626d31e-6b79-44a4-b29d-0912f08ccbc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.078122 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7626d31e-6b79-44a4-b29d-0912f08ccbc7-kube-api-access-htvr9" (OuterVolumeSpecName: "kube-api-access-htvr9") pod "7626d31e-6b79-44a4-b29d-0912f08ccbc7" (UID: "7626d31e-6b79-44a4-b29d-0912f08ccbc7"). InnerVolumeSpecName "kube-api-access-htvr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.100401 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7626d31e-6b79-44a4-b29d-0912f08ccbc7" (UID: "7626d31e-6b79-44a4-b29d-0912f08ccbc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.106161 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-config-data" (OuterVolumeSpecName: "config-data") pod "7626d31e-6b79-44a4-b29d-0912f08ccbc7" (UID: "7626d31e-6b79-44a4-b29d-0912f08ccbc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.162055 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.162097 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htvr9\" (UniqueName: \"kubernetes.io/projected/7626d31e-6b79-44a4-b29d-0912f08ccbc7-kube-api-access-htvr9\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.162116 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.162128 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7626d31e-6b79-44a4-b29d-0912f08ccbc7-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.451688 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfjj4" event={"ID":"f82d7a4d-d310-4f1d-a382-757ee0b5b794","Type":"ContainerStarted","Data":"716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b"} Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.455962 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rnkr2" event={"ID":"7626d31e-6b79-44a4-b29d-0912f08ccbc7","Type":"ContainerDied","Data":"5a5dc9e77581df8709360f12e5b2586e64c62f0fa2a22c08105a04c0d1abd533"} Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.456141 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a5dc9e77581df8709360f12e5b2586e64c62f0fa2a22c08105a04c0d1abd533" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.456035 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rnkr2" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.498834 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hfjj4" podStartSLOduration=2.8543174650000003 podStartE2EDuration="5.498782029s" podCreationTimestamp="2025-10-07 22:32:02 +0000 UTC" firstStartedPulling="2025-10-07 22:32:04.401138906 +0000 UTC m=+1398.203837009" lastFinishedPulling="2025-10-07 22:32:07.04560347 +0000 UTC m=+1400.848301573" observedRunningTime="2025-10-07 22:32:07.477784605 +0000 UTC m=+1401.280482708" watchObservedRunningTime="2025-10-07 22:32:07.498782029 +0000 UTC m=+1401.301480122" Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.675812 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.676233 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerName="nova-api-log" containerID="cri-o://5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec" gracePeriod=30 Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.676249 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerName="nova-api-api" containerID="cri-o://267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d" gracePeriod=30 Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.701475 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.706213 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a1a77ea3-6c18-41ef-a974-b285e03c0c55" containerName="nova-scheduler-scheduler" containerID="cri-o://9b4843e4b148cfbebe1ac315d6231b9c9dee183c546e3c9adaf0cd06cd4cdcaa" gracePeriod=30 Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.764245 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.764591 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-log" containerID="cri-o://05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101" gracePeriod=30 Oct 07 22:32:07 crc kubenswrapper[4871]: I1007 22:32:07.764697 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-metadata" containerID="cri-o://be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e" gracePeriod=30 Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.294472 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.399593 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-config-data\") pod \"03a6b072-112e-4e9d-8f58-669d4127c21b\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.400175 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-internal-tls-certs\") pod \"03a6b072-112e-4e9d-8f58-669d4127c21b\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.400354 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxhtl\" (UniqueName: \"kubernetes.io/projected/03a6b072-112e-4e9d-8f58-669d4127c21b-kube-api-access-mxhtl\") pod \"03a6b072-112e-4e9d-8f58-669d4127c21b\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.400424 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-public-tls-certs\") pod \"03a6b072-112e-4e9d-8f58-669d4127c21b\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.400506 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-combined-ca-bundle\") pod \"03a6b072-112e-4e9d-8f58-669d4127c21b\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.400583 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03a6b072-112e-4e9d-8f58-669d4127c21b-logs\") pod \"03a6b072-112e-4e9d-8f58-669d4127c21b\" (UID: \"03a6b072-112e-4e9d-8f58-669d4127c21b\") " Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.401543 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a6b072-112e-4e9d-8f58-669d4127c21b-logs" (OuterVolumeSpecName: "logs") pod "03a6b072-112e-4e9d-8f58-669d4127c21b" (UID: "03a6b072-112e-4e9d-8f58-669d4127c21b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.417721 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03a6b072-112e-4e9d-8f58-669d4127c21b-kube-api-access-mxhtl" (OuterVolumeSpecName: "kube-api-access-mxhtl") pod "03a6b072-112e-4e9d-8f58-669d4127c21b" (UID: "03a6b072-112e-4e9d-8f58-669d4127c21b"). InnerVolumeSpecName "kube-api-access-mxhtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.445016 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03a6b072-112e-4e9d-8f58-669d4127c21b" (UID: "03a6b072-112e-4e9d-8f58-669d4127c21b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.466419 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "03a6b072-112e-4e9d-8f58-669d4127c21b" (UID: "03a6b072-112e-4e9d-8f58-669d4127c21b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.467954 4871 generic.go:334] "Generic (PLEG): container finished" podID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerID="267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d" exitCode=0 Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.467997 4871 generic.go:334] "Generic (PLEG): container finished" podID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerID="5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec" exitCode=143 Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.468056 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03a6b072-112e-4e9d-8f58-669d4127c21b","Type":"ContainerDied","Data":"267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d"} Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.468090 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03a6b072-112e-4e9d-8f58-669d4127c21b","Type":"ContainerDied","Data":"5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec"} Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.468100 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03a6b072-112e-4e9d-8f58-669d4127c21b","Type":"ContainerDied","Data":"059eb4ebcd4dcc439fcde0c02c2a64e37ec1c4fae9d1c51a9d477f7a6be80bb5"} Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.468118 4871 scope.go:117] "RemoveContainer" containerID="267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.467961 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "03a6b072-112e-4e9d-8f58-669d4127c21b" (UID: "03a6b072-112e-4e9d-8f58-669d4127c21b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.469077 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.471356 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-config-data" (OuterVolumeSpecName: "config-data") pod "03a6b072-112e-4e9d-8f58-669d4127c21b" (UID: "03a6b072-112e-4e9d-8f58-669d4127c21b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.472057 4871 generic.go:334] "Generic (PLEG): container finished" podID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerID="05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101" exitCode=143 Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.472146 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76","Type":"ContainerDied","Data":"05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101"} Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.503745 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03a6b072-112e-4e9d-8f58-669d4127c21b-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.503910 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.504021 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.506541 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxhtl\" (UniqueName: \"kubernetes.io/projected/03a6b072-112e-4e9d-8f58-669d4127c21b-kube-api-access-mxhtl\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.506681 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.506773 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03a6b072-112e-4e9d-8f58-669d4127c21b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.523854 4871 scope.go:117] "RemoveContainer" containerID="5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.555235 4871 scope.go:117] "RemoveContainer" containerID="267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d" Oct 07 22:32:08 crc kubenswrapper[4871]: E1007 22:32:08.555774 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d\": container with ID starting with 267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d not found: ID does not exist" containerID="267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.555827 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d"} err="failed to get container status \"267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d\": rpc error: code = NotFound desc = could not find container \"267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d\": container with ID starting with 267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d not found: ID does not exist" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.555852 4871 scope.go:117] "RemoveContainer" containerID="5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec" Oct 07 22:32:08 crc kubenswrapper[4871]: E1007 22:32:08.556364 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec\": container with ID starting with 5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec not found: ID does not exist" containerID="5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.556384 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec"} err="failed to get container status \"5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec\": rpc error: code = NotFound desc = could not find container \"5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec\": container with ID starting with 5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec not found: ID does not exist" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.556397 4871 scope.go:117] "RemoveContainer" containerID="267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.556604 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d"} err="failed to get container status \"267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d\": rpc error: code = NotFound desc = could not find container \"267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d\": container with ID starting with 267ad890c6823a00c3d910b707341bbecd71307116e1d7b7c3597529a7fe999d not found: ID does not exist" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.556623 4871 scope.go:117] "RemoveContainer" containerID="5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.556847 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec"} err="failed to get container status \"5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec\": rpc error: code = NotFound desc = could not find container \"5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec\": container with ID starting with 5c51d0f06e24484fe61a8e794b655a40991fff523c9c715fff4f746f6f8533ec not found: ID does not exist" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.809772 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.820897 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.833446 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 22:32:08 crc kubenswrapper[4871]: E1007 22:32:08.833906 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerName="nova-api-log" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.833925 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerName="nova-api-log" Oct 07 22:32:08 crc kubenswrapper[4871]: E1007 22:32:08.833952 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerName="nova-api-api" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.833961 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerName="nova-api-api" Oct 07 22:32:08 crc kubenswrapper[4871]: E1007 22:32:08.833975 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7626d31e-6b79-44a4-b29d-0912f08ccbc7" containerName="nova-manage" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.833981 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7626d31e-6b79-44a4-b29d-0912f08ccbc7" containerName="nova-manage" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.834164 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerName="nova-api-api" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.834186 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7626d31e-6b79-44a4-b29d-0912f08ccbc7" containerName="nova-manage" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.834203 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="03a6b072-112e-4e9d-8f58-669d4127c21b" containerName="nova-api-log" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.835249 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.839996 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.840417 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.840488 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.863690 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.914174 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-public-tls-certs\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.914288 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68dde0ec-a90a-4b25-b329-ec645d32671f-logs\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.914334 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.914385 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-config-data\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.914438 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.914512 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxq8l\" (UniqueName: \"kubernetes.io/projected/68dde0ec-a90a-4b25-b329-ec645d32671f-kube-api-access-xxq8l\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:08 crc kubenswrapper[4871]: I1007 22:32:08.994723 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03a6b072-112e-4e9d-8f58-669d4127c21b" path="/var/lib/kubelet/pods/03a6b072-112e-4e9d-8f58-669d4127c21b/volumes" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.017088 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68dde0ec-a90a-4b25-b329-ec645d32671f-logs\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.018134 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.018008 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68dde0ec-a90a-4b25-b329-ec645d32671f-logs\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.018186 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-config-data\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.018330 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.018584 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxq8l\" (UniqueName: \"kubernetes.io/projected/68dde0ec-a90a-4b25-b329-ec645d32671f-kube-api-access-xxq8l\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.018775 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-public-tls-certs\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.024680 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.025472 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-config-data\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.026807 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-public-tls-certs\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.034832 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.039156 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxq8l\" (UniqueName: \"kubernetes.io/projected/68dde0ec-a90a-4b25-b329-ec645d32671f-kube-api-access-xxq8l\") pod \"nova-api-0\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.165105 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.484700 4871 generic.go:334] "Generic (PLEG): container finished" podID="a1a77ea3-6c18-41ef-a974-b285e03c0c55" containerID="9b4843e4b148cfbebe1ac315d6231b9c9dee183c546e3c9adaf0cd06cd4cdcaa" exitCode=0 Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.484919 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a1a77ea3-6c18-41ef-a974-b285e03c0c55","Type":"ContainerDied","Data":"9b4843e4b148cfbebe1ac315d6231b9c9dee183c546e3c9adaf0cd06cd4cdcaa"} Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.667192 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:32:09 crc kubenswrapper[4871]: W1007 22:32:09.677235 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68dde0ec_a90a_4b25_b329_ec645d32671f.slice/crio-61ebb153ce635bf338df4f6177e97ab9c1321079b1e7c75b3d6c18c4f67d2324 WatchSource:0}: Error finding container 61ebb153ce635bf338df4f6177e97ab9c1321079b1e7c75b3d6c18c4f67d2324: Status 404 returned error can't find the container with id 61ebb153ce635bf338df4f6177e97ab9c1321079b1e7c75b3d6c18c4f67d2324 Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.809030 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.939011 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-combined-ca-bundle\") pod \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.939483 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kkql\" (UniqueName: \"kubernetes.io/projected/a1a77ea3-6c18-41ef-a974-b285e03c0c55-kube-api-access-7kkql\") pod \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.939664 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-config-data\") pod \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\" (UID: \"a1a77ea3-6c18-41ef-a974-b285e03c0c55\") " Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.954076 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a77ea3-6c18-41ef-a974-b285e03c0c55-kube-api-access-7kkql" (OuterVolumeSpecName: "kube-api-access-7kkql") pod "a1a77ea3-6c18-41ef-a974-b285e03c0c55" (UID: "a1a77ea3-6c18-41ef-a974-b285e03c0c55"). InnerVolumeSpecName "kube-api-access-7kkql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.986168 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-config-data" (OuterVolumeSpecName: "config-data") pod "a1a77ea3-6c18-41ef-a974-b285e03c0c55" (UID: "a1a77ea3-6c18-41ef-a974-b285e03c0c55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:09 crc kubenswrapper[4871]: I1007 22:32:09.998293 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1a77ea3-6c18-41ef-a974-b285e03c0c55" (UID: "a1a77ea3-6c18-41ef-a974-b285e03c0c55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.043353 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.043478 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a77ea3-6c18-41ef-a974-b285e03c0c55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.043539 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kkql\" (UniqueName: \"kubernetes.io/projected/a1a77ea3-6c18-41ef-a974-b285e03c0c55-kube-api-access-7kkql\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.495189 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.495187 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a1a77ea3-6c18-41ef-a974-b285e03c0c55","Type":"ContainerDied","Data":"2ce05702e7336421aedfb550535beb0bd67cfbd9192a2cc1935c216b66f86b89"} Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.495670 4871 scope.go:117] "RemoveContainer" containerID="9b4843e4b148cfbebe1ac315d6231b9c9dee183c546e3c9adaf0cd06cd4cdcaa" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.500051 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68dde0ec-a90a-4b25-b329-ec645d32671f","Type":"ContainerStarted","Data":"2bf5bc8c358110db521839c0ee5aeb6e45d6fb3a5fe434cba3a0bfc2e3f361c7"} Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.500102 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68dde0ec-a90a-4b25-b329-ec645d32671f","Type":"ContainerStarted","Data":"893dcbed9540fca5394cb10bce40d68accf0e792ccb53a0bca71eeb8cf8337c3"} Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.500113 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68dde0ec-a90a-4b25-b329-ec645d32671f","Type":"ContainerStarted","Data":"61ebb153ce635bf338df4f6177e97ab9c1321079b1e7c75b3d6c18c4f67d2324"} Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.519122 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.5190961339999998 podStartE2EDuration="2.519096134s" podCreationTimestamp="2025-10-07 22:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:32:10.51858585 +0000 UTC m=+1404.321283923" watchObservedRunningTime="2025-10-07 22:32:10.519096134 +0000 UTC m=+1404.321794197" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.543405 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.554271 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.566010 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:32:10 crc kubenswrapper[4871]: E1007 22:32:10.566409 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1a77ea3-6c18-41ef-a974-b285e03c0c55" containerName="nova-scheduler-scheduler" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.566426 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1a77ea3-6c18-41ef-a974-b285e03c0c55" containerName="nova-scheduler-scheduler" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.566649 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1a77ea3-6c18-41ef-a974-b285e03c0c55" containerName="nova-scheduler-scheduler" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.567319 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.569737 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.588897 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.656054 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-config-data\") pod \"nova-scheduler-0\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.656180 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcmfw\" (UniqueName: \"kubernetes.io/projected/ab7d1def-9358-44f0-9245-e333fb60bac4-kube-api-access-rcmfw\") pod \"nova-scheduler-0\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.656287 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.758210 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-config-data\") pod \"nova-scheduler-0\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.758304 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcmfw\" (UniqueName: \"kubernetes.io/projected/ab7d1def-9358-44f0-9245-e333fb60bac4-kube-api-access-rcmfw\") pod \"nova-scheduler-0\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.758376 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.762701 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.762781 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-config-data\") pod \"nova-scheduler-0\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.779063 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcmfw\" (UniqueName: \"kubernetes.io/projected/ab7d1def-9358-44f0-9245-e333fb60bac4-kube-api-access-rcmfw\") pod \"nova-scheduler-0\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.888019 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.915579 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:34036->10.217.0.192:8775: read: connection reset by peer" Oct 07 22:32:10 crc kubenswrapper[4871]: I1007 22:32:10.915602 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:34038->10.217.0.192:8775: read: connection reset by peer" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.001368 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1a77ea3-6c18-41ef-a974-b285e03c0c55" path="/var/lib/kubelet/pods/a1a77ea3-6c18-41ef-a974-b285e03c0c55/volumes" Oct 07 22:32:11 crc kubenswrapper[4871]: W1007 22:32:11.425111 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab7d1def_9358_44f0_9245_e333fb60bac4.slice/crio-216394cc8d96b3a3aa867cf7ad54dd4656d2395398530a696359975606a58fc1 WatchSource:0}: Error finding container 216394cc8d96b3a3aa867cf7ad54dd4656d2395398530a696359975606a58fc1: Status 404 returned error can't find the container with id 216394cc8d96b3a3aa867cf7ad54dd4656d2395398530a696359975606a58fc1 Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.425831 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.492409 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.519216 4871 generic.go:334] "Generic (PLEG): container finished" podID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerID="be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e" exitCode=0 Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.519322 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76","Type":"ContainerDied","Data":"be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e"} Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.519350 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.519390 4871 scope.go:117] "RemoveContainer" containerID="be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.519366 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76","Type":"ContainerDied","Data":"5103d16c980f7aeff2b3577ef0e68425c243f15606ca5b1f36abfce357b90a51"} Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.526959 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ab7d1def-9358-44f0-9245-e333fb60bac4","Type":"ContainerStarted","Data":"216394cc8d96b3a3aa867cf7ad54dd4656d2395398530a696359975606a58fc1"} Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.571134 4871 scope.go:117] "RemoveContainer" containerID="05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.579460 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-nova-metadata-tls-certs\") pod \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.579661 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-combined-ca-bundle\") pod \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.579844 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq9rk\" (UniqueName: \"kubernetes.io/projected/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-kube-api-access-kq9rk\") pod \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.579874 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-config-data\") pod \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.579908 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-logs\") pod \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\" (UID: \"593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76\") " Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.582245 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-logs" (OuterVolumeSpecName: "logs") pod "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" (UID: "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.587902 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-kube-api-access-kq9rk" (OuterVolumeSpecName: "kube-api-access-kq9rk") pod "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" (UID: "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76"). InnerVolumeSpecName "kube-api-access-kq9rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.610244 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-config-data" (OuterVolumeSpecName: "config-data") pod "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" (UID: "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.616776 4871 scope.go:117] "RemoveContainer" containerID="be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e" Oct 07 22:32:11 crc kubenswrapper[4871]: E1007 22:32:11.618416 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e\": container with ID starting with be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e not found: ID does not exist" containerID="be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.618461 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e"} err="failed to get container status \"be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e\": rpc error: code = NotFound desc = could not find container \"be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e\": container with ID starting with be2e26f5eadefc14bd266aeb466247b352259500c1b495d40984c6e6f394653e not found: ID does not exist" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.618504 4871 scope.go:117] "RemoveContainer" containerID="05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101" Oct 07 22:32:11 crc kubenswrapper[4871]: E1007 22:32:11.618978 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101\": container with ID starting with 05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101 not found: ID does not exist" containerID="05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.619002 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101"} err="failed to get container status \"05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101\": rpc error: code = NotFound desc = could not find container \"05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101\": container with ID starting with 05ffe8e372554cb939304f01ae15b5a20c3f244963d7e93bc41f74c3240f7101 not found: ID does not exist" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.626121 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" (UID: "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.644753 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" (UID: "593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.682643 4871 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.682691 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.682704 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq9rk\" (UniqueName: \"kubernetes.io/projected/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-kube-api-access-kq9rk\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.682714 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.682723 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.851402 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.858549 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.876525 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:32:11 crc kubenswrapper[4871]: E1007 22:32:11.889048 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-metadata" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.889105 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-metadata" Oct 07 22:32:11 crc kubenswrapper[4871]: E1007 22:32:11.889144 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-log" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.889157 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-log" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.889674 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-metadata" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.889706 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" containerName="nova-metadata-log" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.890941 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.891064 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.894336 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.894507 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.988471 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-config-data\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.988615 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b48c5c-8fd6-4578-bc71-3f7b47860649-logs\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.988700 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgf2v\" (UniqueName: \"kubernetes.io/projected/23b48c5c-8fd6-4578-bc71-3f7b47860649-kube-api-access-xgf2v\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.988851 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:11 crc kubenswrapper[4871]: I1007 22:32:11.988901 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.090520 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-config-data\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.090571 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b48c5c-8fd6-4578-bc71-3f7b47860649-logs\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.090645 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgf2v\" (UniqueName: \"kubernetes.io/projected/23b48c5c-8fd6-4578-bc71-3f7b47860649-kube-api-access-xgf2v\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.090768 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.090830 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.092349 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b48c5c-8fd6-4578-bc71-3f7b47860649-logs\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.096978 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-config-data\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.098255 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.103345 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.111784 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgf2v\" (UniqueName: \"kubernetes.io/projected/23b48c5c-8fd6-4578-bc71-3f7b47860649-kube-api-access-xgf2v\") pod \"nova-metadata-0\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.282135 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.556939 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ab7d1def-9358-44f0-9245-e333fb60bac4","Type":"ContainerStarted","Data":"d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a"} Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.585604 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.585572962 podStartE2EDuration="2.585572962s" podCreationTimestamp="2025-10-07 22:32:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:32:12.580896499 +0000 UTC m=+1406.383594582" watchObservedRunningTime="2025-10-07 22:32:12.585572962 +0000 UTC m=+1406.388271055" Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.812905 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:32:12 crc kubenswrapper[4871]: W1007 22:32:12.817218 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23b48c5c_8fd6_4578_bc71_3f7b47860649.slice/crio-03f7e7bbcdc1cbea796f537923a2f519f4899845d8d312992fcd10ea14e3d1dd WatchSource:0}: Error finding container 03f7e7bbcdc1cbea796f537923a2f519f4899845d8d312992fcd10ea14e3d1dd: Status 404 returned error can't find the container with id 03f7e7bbcdc1cbea796f537923a2f519f4899845d8d312992fcd10ea14e3d1dd Oct 07 22:32:12 crc kubenswrapper[4871]: I1007 22:32:12.997474 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76" path="/var/lib/kubelet/pods/593b8c4c-db88-4fdd-8a40-a3a2ea7a3b76/volumes" Oct 07 22:32:13 crc kubenswrapper[4871]: I1007 22:32:13.162420 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:13 crc kubenswrapper[4871]: I1007 22:32:13.162512 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:13 crc kubenswrapper[4871]: I1007 22:32:13.232247 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:13 crc kubenswrapper[4871]: I1007 22:32:13.585132 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23b48c5c-8fd6-4578-bc71-3f7b47860649","Type":"ContainerStarted","Data":"83a8879c5c8652cd944b767603633f0434212ef09dde9ebec772c3c1fe6093a1"} Oct 07 22:32:13 crc kubenswrapper[4871]: I1007 22:32:13.585977 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23b48c5c-8fd6-4578-bc71-3f7b47860649","Type":"ContainerStarted","Data":"0e1f568994d1e614f92908e5ccc4ef98ce4c080d84575da81deff6e86b817516"} Oct 07 22:32:13 crc kubenswrapper[4871]: I1007 22:32:13.586024 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23b48c5c-8fd6-4578-bc71-3f7b47860649","Type":"ContainerStarted","Data":"03f7e7bbcdc1cbea796f537923a2f519f4899845d8d312992fcd10ea14e3d1dd"} Oct 07 22:32:13 crc kubenswrapper[4871]: I1007 22:32:13.677231 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:13 crc kubenswrapper[4871]: I1007 22:32:13.715367 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.715341154 podStartE2EDuration="2.715341154s" podCreationTimestamp="2025-10-07 22:32:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:32:13.615048893 +0000 UTC m=+1407.417746976" watchObservedRunningTime="2025-10-07 22:32:13.715341154 +0000 UTC m=+1407.518039227" Oct 07 22:32:13 crc kubenswrapper[4871]: I1007 22:32:13.754700 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hfjj4"] Oct 07 22:32:15 crc kubenswrapper[4871]: I1007 22:32:15.604578 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hfjj4" podUID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerName="registry-server" containerID="cri-o://716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b" gracePeriod=2 Oct 07 22:32:15 crc kubenswrapper[4871]: I1007 22:32:15.889225 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.147947 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.208935 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t26v\" (UniqueName: \"kubernetes.io/projected/f82d7a4d-d310-4f1d-a382-757ee0b5b794-kube-api-access-6t26v\") pod \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.209924 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-utilities\") pod \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.210771 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-utilities" (OuterVolumeSpecName: "utilities") pod "f82d7a4d-d310-4f1d-a382-757ee0b5b794" (UID: "f82d7a4d-d310-4f1d-a382-757ee0b5b794"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.210935 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-catalog-content\") pod \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\" (UID: \"f82d7a4d-d310-4f1d-a382-757ee0b5b794\") " Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.213362 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.220095 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f82d7a4d-d310-4f1d-a382-757ee0b5b794-kube-api-access-6t26v" (OuterVolumeSpecName: "kube-api-access-6t26v") pod "f82d7a4d-d310-4f1d-a382-757ee0b5b794" (UID: "f82d7a4d-d310-4f1d-a382-757ee0b5b794"). InnerVolumeSpecName "kube-api-access-6t26v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.315627 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t26v\" (UniqueName: \"kubernetes.io/projected/f82d7a4d-d310-4f1d-a382-757ee0b5b794-kube-api-access-6t26v\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.616192 4871 generic.go:334] "Generic (PLEG): container finished" podID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerID="716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b" exitCode=0 Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.616268 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfjj4" event={"ID":"f82d7a4d-d310-4f1d-a382-757ee0b5b794","Type":"ContainerDied","Data":"716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b"} Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.616317 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfjj4" event={"ID":"f82d7a4d-d310-4f1d-a382-757ee0b5b794","Type":"ContainerDied","Data":"c9eae584ffd66e6de840c57434c9533541750b6f7c06b5f61b0d6ddf36466984"} Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.616343 4871 scope.go:117] "RemoveContainer" containerID="716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.616402 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfjj4" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.645681 4871 scope.go:117] "RemoveContainer" containerID="59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.674645 4871 scope.go:117] "RemoveContainer" containerID="ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.690656 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f82d7a4d-d310-4f1d-a382-757ee0b5b794" (UID: "f82d7a4d-d310-4f1d-a382-757ee0b5b794"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.726603 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82d7a4d-d310-4f1d-a382-757ee0b5b794-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.730322 4871 scope.go:117] "RemoveContainer" containerID="716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b" Oct 07 22:32:16 crc kubenswrapper[4871]: E1007 22:32:16.731268 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b\": container with ID starting with 716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b not found: ID does not exist" containerID="716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.731366 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b"} err="failed to get container status \"716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b\": rpc error: code = NotFound desc = could not find container \"716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b\": container with ID starting with 716e5c7ecfb6d108404ea913821ae427c02b76004ad64f6aeb6fc1ea644e348b not found: ID does not exist" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.731418 4871 scope.go:117] "RemoveContainer" containerID="59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45" Oct 07 22:32:16 crc kubenswrapper[4871]: E1007 22:32:16.731899 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45\": container with ID starting with 59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45 not found: ID does not exist" containerID="59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.731955 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45"} err="failed to get container status \"59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45\": rpc error: code = NotFound desc = could not find container \"59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45\": container with ID starting with 59806ed5c66f1841005bffa6884baf97b185182f351927b8839b5811601ecd45 not found: ID does not exist" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.731997 4871 scope.go:117] "RemoveContainer" containerID="ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe" Oct 07 22:32:16 crc kubenswrapper[4871]: E1007 22:32:16.732410 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe\": container with ID starting with ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe not found: ID does not exist" containerID="ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.732458 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe"} err="failed to get container status \"ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe\": rpc error: code = NotFound desc = could not find container \"ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe\": container with ID starting with ba60172a544f2a3d3178c08caeb85a7778823047fbc6be30e1d6b3e7596b39fe not found: ID does not exist" Oct 07 22:32:16 crc kubenswrapper[4871]: I1007 22:32:16.976709 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hfjj4"] Oct 07 22:32:17 crc kubenswrapper[4871]: I1007 22:32:17.008033 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hfjj4"] Oct 07 22:32:17 crc kubenswrapper[4871]: I1007 22:32:17.283210 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 22:32:17 crc kubenswrapper[4871]: I1007 22:32:17.283282 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 22:32:18 crc kubenswrapper[4871]: I1007 22:32:18.998028 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" path="/var/lib/kubelet/pods/f82d7a4d-d310-4f1d-a382-757ee0b5b794/volumes" Oct 07 22:32:19 crc kubenswrapper[4871]: I1007 22:32:19.165663 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 22:32:19 crc kubenswrapper[4871]: I1007 22:32:19.165756 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 22:32:20 crc kubenswrapper[4871]: I1007 22:32:20.186946 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 22:32:20 crc kubenswrapper[4871]: I1007 22:32:20.186996 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 22:32:20 crc kubenswrapper[4871]: I1007 22:32:20.888656 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 22:32:20 crc kubenswrapper[4871]: I1007 22:32:20.928646 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 22:32:21 crc kubenswrapper[4871]: I1007 22:32:21.723678 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 22:32:22 crc kubenswrapper[4871]: I1007 22:32:22.284057 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 22:32:22 crc kubenswrapper[4871]: I1007 22:32:22.284175 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 22:32:23 crc kubenswrapper[4871]: I1007 22:32:23.302120 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 22:32:23 crc kubenswrapper[4871]: I1007 22:32:23.302151 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 22:32:27 crc kubenswrapper[4871]: I1007 22:32:27.795047 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 22:32:29 crc kubenswrapper[4871]: I1007 22:32:29.177157 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 22:32:29 crc kubenswrapper[4871]: I1007 22:32:29.178359 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 22:32:29 crc kubenswrapper[4871]: I1007 22:32:29.180160 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 22:32:29 crc kubenswrapper[4871]: I1007 22:32:29.197884 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 22:32:29 crc kubenswrapper[4871]: I1007 22:32:29.769262 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 22:32:29 crc kubenswrapper[4871]: I1007 22:32:29.780077 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 22:32:32 crc kubenswrapper[4871]: I1007 22:32:32.292963 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 22:32:32 crc kubenswrapper[4871]: I1007 22:32:32.295475 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 22:32:32 crc kubenswrapper[4871]: I1007 22:32:32.300213 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 22:32:32 crc kubenswrapper[4871]: I1007 22:32:32.811896 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.621842 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.622781 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="943e025c-54d8-4686-81ee-894259d7af56" containerName="openstackclient" containerID="cri-o://0bf2fc0c8cbcc24b03cb64efbb163db4f05f24420ea760d649a0be66f64959ec" gracePeriod=2 Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.685611 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.840417 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.840708 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="ovn-northd" containerID="cri-o://691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303" gracePeriod=30 Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.841199 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="openstack-network-exporter" containerID="cri-o://a0bcaf69c69657ecf8f36f03f56417c9e0e10c4b3d9151b2ed50c569b93152aa" gracePeriod=30 Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.958501 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinderf74a-account-delete-mvrf4"] Oct 07 22:32:51 crc kubenswrapper[4871]: E1007 22:32:51.959094 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerName="extract-utilities" Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.959117 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerName="extract-utilities" Oct 07 22:32:51 crc kubenswrapper[4871]: E1007 22:32:51.959148 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerName="registry-server" Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.959155 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerName="registry-server" Oct 07 22:32:51 crc kubenswrapper[4871]: E1007 22:32:51.959172 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerName="extract-content" Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.959179 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerName="extract-content" Oct 07 22:32:51 crc kubenswrapper[4871]: E1007 22:32:51.959192 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="943e025c-54d8-4686-81ee-894259d7af56" containerName="openstackclient" Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.959198 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="943e025c-54d8-4686-81ee-894259d7af56" containerName="openstackclient" Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.959430 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="943e025c-54d8-4686-81ee-894259d7af56" containerName="openstackclient" Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.959447 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f82d7a4d-d310-4f1d-a382-757ee0b5b794" containerName="registry-server" Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.974524 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderf74a-account-delete-mvrf4" Oct 07 22:32:51 crc kubenswrapper[4871]: I1007 22:32:51.996546 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderf74a-account-delete-mvrf4"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.005455 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkrz5\" (UniqueName: \"kubernetes.io/projected/1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10-kube-api-access-nkrz5\") pod \"cinderf74a-account-delete-mvrf4\" (UID: \"1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10\") " pod="openstack/cinderf74a-account-delete-mvrf4" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.005655 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.006126 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerName="openstack-network-exporter" containerID="cri-o://d94cfb46e87419008029b8ee502b157dab2ee74b35e86d31ffb0102836fd1648" gracePeriod=300 Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.038510 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glancea79e-account-delete-jv6lz"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.040001 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancea79e-account-delete-jv6lz" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.056939 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancea79e-account-delete-jv6lz"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.134437 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.149388 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdjq5\" (UniqueName: \"kubernetes.io/projected/b17ba7e1-deb0-4a31-8874-ee1e6686a3f8-kube-api-access-wdjq5\") pod \"glancea79e-account-delete-jv6lz\" (UID: \"b17ba7e1-deb0-4a31-8874-ee1e6686a3f8\") " pod="openstack/glancea79e-account-delete-jv6lz" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.150146 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkrz5\" (UniqueName: \"kubernetes.io/projected/1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10-kube-api-access-nkrz5\") pod \"cinderf74a-account-delete-mvrf4\" (UID: \"1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10\") " pod="openstack/cinderf74a-account-delete-mvrf4" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.179330 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerName="ovsdbserver-sb" containerID="cri-o://99e861992db1f9c7304f7a584cbddbe67197ed224df105828d0b35bd4e191a37" gracePeriod=300 Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.246621 4871 generic.go:334] "Generic (PLEG): container finished" podID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerID="a0bcaf69c69657ecf8f36f03f56417c9e0e10c4b3d9151b2ed50c569b93152aa" exitCode=2 Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.249731 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fb7d0be2-8b44-44ed-948d-11e1932d27b4","Type":"ContainerDied","Data":"a0bcaf69c69657ecf8f36f03f56417c9e0e10c4b3d9151b2ed50c569b93152aa"} Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.355190 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkrz5\" (UniqueName: \"kubernetes.io/projected/1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10-kube-api-access-nkrz5\") pod \"cinderf74a-account-delete-mvrf4\" (UID: \"1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10\") " pod="openstack/cinderf74a-account-delete-mvrf4" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.375643 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdjq5\" (UniqueName: \"kubernetes.io/projected/b17ba7e1-deb0-4a31-8874-ee1e6686a3f8-kube-api-access-wdjq5\") pod \"glancea79e-account-delete-jv6lz\" (UID: \"b17ba7e1-deb0-4a31-8874-ee1e6686a3f8\") " pod="openstack/glancea79e-account-delete-jv6lz" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.376664 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementf2b9-account-delete-llr4x"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.378483 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementf2b9-account-delete-llr4x" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.405697 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderf74a-account-delete-mvrf4" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.447046 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdjq5\" (UniqueName: \"kubernetes.io/projected/b17ba7e1-deb0-4a31-8874-ee1e6686a3f8-kube-api-access-wdjq5\") pod \"glancea79e-account-delete-jv6lz\" (UID: \"b17ba7e1-deb0-4a31-8874-ee1e6686a3f8\") " pod="openstack/glancea79e-account-delete-jv6lz" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.467882 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementf2b9-account-delete-llr4x"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.482222 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9wcx\" (UniqueName: \"kubernetes.io/projected/361d6d85-f8c7-4643-95c4-29741f91a179-kube-api-access-r9wcx\") pod \"placementf2b9-account-delete-llr4x\" (UID: \"361d6d85-f8c7-4643-95c4-29741f91a179\") " pod="openstack/placementf2b9-account-delete-llr4x" Oct 07 22:32:52 crc kubenswrapper[4871]: E1007 22:32:52.484553 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 07 22:32:52 crc kubenswrapper[4871]: E1007 22:32:52.484650 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data podName:dd77ac8e-36b6-4e16-869a-3ce6bff223ac nodeName:}" failed. No retries permitted until 2025-10-07 22:32:52.984619729 +0000 UTC m=+1446.787317802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data") pod "rabbitmq-server-0" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac") : configmap "rabbitmq-config-data" not found Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.523451 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancea79e-account-delete-jv6lz" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.590015 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9wcx\" (UniqueName: \"kubernetes.io/projected/361d6d85-f8c7-4643-95c4-29741f91a179-kube-api-access-r9wcx\") pod \"placementf2b9-account-delete-llr4x\" (UID: \"361d6d85-f8c7-4643-95c4-29741f91a179\") " pod="openstack/placementf2b9-account-delete-llr4x" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.607983 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7sq86"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.645510 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9wcx\" (UniqueName: \"kubernetes.io/projected/361d6d85-f8c7-4643-95c4-29741f91a179-kube-api-access-r9wcx\") pod \"placementf2b9-account-delete-llr4x\" (UID: \"361d6d85-f8c7-4643-95c4-29741f91a179\") " pod="openstack/placementf2b9-account-delete-llr4x" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.647648 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7sq86"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.673672 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapid46f-account-delete-8z75z"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.675300 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapid46f-account-delete-8z75z" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.688227 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapid46f-account-delete-8z75z"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.729856 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-tkjl5"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.780033 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-tkjl5"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.807644 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9l76\" (UniqueName: \"kubernetes.io/projected/ca98e71f-5147-4bb7-bd76-e4bb11cfcf42-kube-api-access-q9l76\") pod \"novaapid46f-account-delete-8z75z\" (UID: \"ca98e71f-5147-4bb7-bd76-e4bb11cfcf42\") " pod="openstack/novaapid46f-account-delete-8z75z" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.822113 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementf2b9-account-delete-llr4x" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.880133 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zt6f9"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.910832 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9l76\" (UniqueName: \"kubernetes.io/projected/ca98e71f-5147-4bb7-bd76-e4bb11cfcf42-kube-api-access-q9l76\") pod \"novaapid46f-account-delete-8z75z\" (UID: \"ca98e71f-5147-4bb7-bd76-e4bb11cfcf42\") " pod="openstack/novaapid46f-account-delete-8z75z" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.917321 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zt6f9"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.945756 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-gz6cr"] Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.946423 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9l76\" (UniqueName: \"kubernetes.io/projected/ca98e71f-5147-4bb7-bd76-e4bb11cfcf42-kube-api-access-q9l76\") pod \"novaapid46f-account-delete-8z75z\" (UID: \"ca98e71f-5147-4bb7-bd76-e4bb11cfcf42\") " pod="openstack/novaapid46f-account-delete-8z75z" Oct 07 22:32:52 crc kubenswrapper[4871]: I1007 22:32:52.964173 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-gz6cr"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.003132 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="945eb664-b114-4ff5-bed5-cfdaa9a99939" path="/var/lib/kubelet/pods/945eb664-b114-4ff5-bed5-cfdaa9a99939/volumes" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.003689 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6" path="/var/lib/kubelet/pods/a4d9fa8a-5dda-4c94-a77c-cce3e73dd9c6/volumes" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.004717 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d464df0c-28ef-4fdd-b70c-307e736def2a" path="/var/lib/kubelet/pods/d464df0c-28ef-4fdd-b70c-307e736def2a/volumes" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.007089 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0b7dc36-3f82-4650-b926-03de75dda7e4" path="/var/lib/kubelet/pods/f0b7dc36-3f82-4650-b926-03de75dda7e4/volumes" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.007679 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-rnkr2"] Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.041576 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.041707 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data podName:dd77ac8e-36b6-4e16-869a-3ce6bff223ac nodeName:}" failed. No retries permitted until 2025-10-07 22:32:54.041665689 +0000 UTC m=+1447.844363762 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data") pod "rabbitmq-server-0" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac") : configmap "rabbitmq-config-data" not found Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.048932 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapid46f-account-delete-8z75z" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.080423 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-rnkr2"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.102085 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ptrx5"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.118740 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ptrx5"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.138975 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8vjd9"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.151947 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.196865 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-vxlsj"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.197182 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-vxlsj" podUID="ec7b131b-1100-4f58-bf1c-012ab142f03b" containerName="openstack-network-exporter" containerID="cri-o://0dce78a091797582beabb0faed2c660e6204b78527df1d3a95b8d0c04c1f1d92" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.242148 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-jhh4r"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.256900 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-7zpls"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.272520 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-7zpls"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.279894 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0a535bdf-4211-4f42-8137-ee421d6a7fae/ovsdbserver-sb/0.log" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.279961 4871 generic.go:334] "Generic (PLEG): container finished" podID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerID="d94cfb46e87419008029b8ee502b157dab2ee74b35e86d31ffb0102836fd1648" exitCode=2 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.287136 4871 generic.go:334] "Generic (PLEG): container finished" podID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerID="99e861992db1f9c7304f7a584cbddbe67197ed224df105828d0b35bd4e191a37" exitCode=143 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.308069 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0a535bdf-4211-4f42-8137-ee421d6a7fae","Type":"ContainerDied","Data":"d94cfb46e87419008029b8ee502b157dab2ee74b35e86d31ffb0102836fd1648"} Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.308175 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0a535bdf-4211-4f42-8137-ee421d6a7fae","Type":"ContainerDied","Data":"99e861992db1f9c7304f7a584cbddbe67197ed224df105828d0b35bd4e191a37"} Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.340849 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.340928 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data podName:1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb nodeName:}" failed. No retries permitted until 2025-10-07 22:32:53.840909106 +0000 UTC m=+1447.643607169 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data") pod "rabbitmq-cell1-server-0" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb") : configmap "rabbitmq-cell1-config-data" not found Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.341381 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a535bdf_4211_4f42_8137_ee421d6a7fae.slice/crio-99e861992db1f9c7304f7a584cbddbe67197ed224df105828d0b35bd4e191a37.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a535bdf_4211_4f42_8137_ee421d6a7fae.slice/crio-conmon-99e861992db1f9c7304f7a584cbddbe67197ed224df105828d0b35bd4e191a37.scope\": RecentStats: unable to find data in memory cache]" Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.375506 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.390829 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.403731 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.410477 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="f9692529-979b-407e-9826-d93875feda52" containerName="openstack-network-exporter" containerID="cri-o://e287d0f45913ff5f5dd60da81232d17cb6a0dd0bdcea624cee7faad1f4925c41" gracePeriod=300 Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.430209 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.430285 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="ovn-northd" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.453497 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4lp7t"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.521051 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="f9692529-979b-407e-9826-d93875feda52" containerName="ovsdbserver-nb" containerID="cri-o://d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6" gracePeriod=300 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.529114 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4lp7t"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.628428 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.629893 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="825fba74-f028-4445-b041-465ccbbad763" containerName="cinder-scheduler" containerID="cri-o://ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.630988 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="825fba74-f028-4445-b041-465ccbbad763" containerName="probe" containerID="cri-o://b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.714084 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.718359 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" containerName="cinder-api-log" containerID="cri-o://8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.718900 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" containerName="cinder-api" containerID="cri-o://b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.728286 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.733505 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-server" containerID="cri-o://2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.734268 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-server" containerID="cri-o://e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.734724 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="swift-recon-cron" containerID="cri-o://d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.734773 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="rsync" containerID="cri-o://1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.734932 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-expirer" containerID="cri-o://5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.734992 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-updater" containerID="cri-o://2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.735035 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-auditor" containerID="cri-o://1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.735091 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-replicator" containerID="cri-o://8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.735160 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-auditor" containerID="cri-o://e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.735220 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-reaper" containerID="cri-o://ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.735199 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-server" containerID="cri-o://a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.735296 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-auditor" containerID="cri-o://26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.735344 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-updater" containerID="cri-o://d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.735395 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-replicator" containerID="cri-o://e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.735397 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-replicator" containerID="cri-o://35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.759850 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.764402 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerName="glance-log" containerID="cri-o://303e52fe41cb9eb2c55a65b8e6fb2c26f6e89c632f0d230eb1f5ba71eb482fee" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.765942 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerName="glance-httpd" containerID="cri-o://91eb51510e831d42f41a70585ae844bb55e88150c93eae1571a50da925c20bf3" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.816975 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0a535bdf-4211-4f42-8137-ee421d6a7fae/ovsdbserver-sb/0.log" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.817076 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.851244 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7567f46b5d-kb4h9"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.851656 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7567f46b5d-kb4h9" podUID="878958ac-64b8-434b-88ee-f23d3c486765" containerName="placement-log" containerID="cri-o://ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.851859 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7567f46b5d-kb4h9" podUID="878958ac-64b8-434b-88ee-f23d3c486765" containerName="placement-api" containerID="cri-o://858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.866898 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-phtsg"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.869943 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" podUID="e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" containerName="dnsmasq-dns" containerID="cri-o://36507cab7159abb01d536c353f20262fe1f5bacd4aef76bb370e449a8777f532" gracePeriod=10 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.875614 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84wtx\" (UniqueName: \"kubernetes.io/projected/0a535bdf-4211-4f42-8137-ee421d6a7fae-kube-api-access-84wtx\") pod \"0a535bdf-4211-4f42-8137-ee421d6a7fae\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.875727 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-combined-ca-bundle\") pod \"0a535bdf-4211-4f42-8137-ee421d6a7fae\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.875806 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-config\") pod \"0a535bdf-4211-4f42-8137-ee421d6a7fae\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.875843 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"0a535bdf-4211-4f42-8137-ee421d6a7fae\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.875886 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdbserver-sb-tls-certs\") pod \"0a535bdf-4211-4f42-8137-ee421d6a7fae\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.875979 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-scripts\") pod \"0a535bdf-4211-4f42-8137-ee421d6a7fae\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.876079 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdb-rundir\") pod \"0a535bdf-4211-4f42-8137-ee421d6a7fae\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.876133 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-metrics-certs-tls-certs\") pod \"0a535bdf-4211-4f42-8137-ee421d6a7fae\" (UID: \"0a535bdf-4211-4f42-8137-ee421d6a7fae\") " Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.879745 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-config" (OuterVolumeSpecName: "config") pod "0a535bdf-4211-4f42-8137-ee421d6a7fae" (UID: "0a535bdf-4211-4f42-8137-ee421d6a7fae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.880284 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-scripts" (OuterVolumeSpecName: "scripts") pod "0a535bdf-4211-4f42-8137-ee421d6a7fae" (UID: "0a535bdf-4211-4f42-8137-ee421d6a7fae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.880839 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "0a535bdf-4211-4f42-8137-ee421d6a7fae" (UID: "0a535bdf-4211-4f42-8137-ee421d6a7fae"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.881282 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.881304 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a535bdf-4211-4f42-8137-ee421d6a7fae-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.881314 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.881411 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 07 22:32:53 crc kubenswrapper[4871]: E1007 22:32:53.881466 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data podName:1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb nodeName:}" failed. No retries permitted until 2025-10-07 22:32:54.881449522 +0000 UTC m=+1448.684147595 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data") pod "rabbitmq-cell1-server-0" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb") : configmap "rabbitmq-cell1-config-data" not found Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.906470 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.906859 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerName="glance-log" containerID="cri-o://0b297eaadbd30609ad30865b36ac820cfdedfdd27b2cf2f2a38dea46f5c786d9" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.907091 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerName="glance-httpd" containerID="cri-o://a6bcbbe481eb19cf659465c0d4fcb0fa01acc7646b80f2eaea5e2ac2187ebf78" gracePeriod=30 Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.911767 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a535bdf-4211-4f42-8137-ee421d6a7fae-kube-api-access-84wtx" (OuterVolumeSpecName: "kube-api-access-84wtx") pod "0a535bdf-4211-4f42-8137-ee421d6a7fae" (UID: "0a535bdf-4211-4f42-8137-ee421d6a7fae"). InnerVolumeSpecName "kube-api-access-84wtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.921084 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "0a535bdf-4211-4f42-8137-ee421d6a7fae" (UID: "0a535bdf-4211-4f42-8137-ee421d6a7fae"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.935936 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 22:32:53 crc kubenswrapper[4871]: I1007 22:32:53.998256 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.001442 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-log" containerID="cri-o://893dcbed9540fca5394cb10bce40d68accf0e792ccb53a0bca71eeb8cf8337c3" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.003483 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84wtx\" (UniqueName: \"kubernetes.io/projected/0a535bdf-4211-4f42-8137-ee421d6a7fae-kube-api-access-84wtx\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.003518 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.004638 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-api" containerID="cri-o://2bf5bc8c358110db521839c0ee5aeb6e45d6fb3a5fe434cba3a0bfc2e3f361c7" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.028132 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.028563 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-log" containerID="cri-o://0e1f568994d1e614f92908e5ccc4ef98ce4c080d84575da81deff6e86b817516" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.030032 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-metadata" containerID="cri-o://83a8879c5c8652cd944b767603633f0434212ef09dde9ebec772c3c1fe6093a1" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.041352 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-6bxbc"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.052884 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-6bxbc"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.065940 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a535bdf-4211-4f42-8137-ee421d6a7fae" (UID: "0a535bdf-4211-4f42-8137-ee421d6a7fae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.068879 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ae9f-account-create-pqnwn"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.108152 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.123118 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.123245 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data podName:dd77ac8e-36b6-4e16-869a-3ce6bff223ac nodeName:}" failed. No retries permitted until 2025-10-07 22:32:56.123203954 +0000 UTC m=+1449.925902027 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data") pod "rabbitmq-server-0" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac") : configmap "rabbitmq-config-data" not found Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.172928 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ae9f-account-create-pqnwn"] Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.175130 4871 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 07 22:32:54 crc kubenswrapper[4871]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 07 22:32:54 crc kubenswrapper[4871]: + source /usr/local/bin/container-scripts/functions Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNBridge=br-int Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNRemote=tcp:localhost:6642 Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNEncapType=geneve Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNAvailabilityZones= Oct 07 22:32:54 crc kubenswrapper[4871]: ++ EnableChassisAsGateway=true Oct 07 22:32:54 crc kubenswrapper[4871]: ++ PhysicalNetworks= Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNHostName= Oct 07 22:32:54 crc kubenswrapper[4871]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 07 22:32:54 crc kubenswrapper[4871]: ++ ovs_dir=/var/lib/openvswitch Oct 07 22:32:54 crc kubenswrapper[4871]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 07 22:32:54 crc kubenswrapper[4871]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 07 22:32:54 crc kubenswrapper[4871]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 07 22:32:54 crc kubenswrapper[4871]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 07 22:32:54 crc kubenswrapper[4871]: + sleep 0.5 Oct 07 22:32:54 crc kubenswrapper[4871]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 07 22:32:54 crc kubenswrapper[4871]: + cleanup_ovsdb_server_semaphore Oct 07 22:32:54 crc kubenswrapper[4871]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 07 22:32:54 crc kubenswrapper[4871]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 07 22:32:54 crc kubenswrapper[4871]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-jhh4r" message=< Oct 07 22:32:54 crc kubenswrapper[4871]: Exiting ovsdb-server (5) [ OK ] Oct 07 22:32:54 crc kubenswrapper[4871]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 07 22:32:54 crc kubenswrapper[4871]: + source /usr/local/bin/container-scripts/functions Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNBridge=br-int Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNRemote=tcp:localhost:6642 Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNEncapType=geneve Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNAvailabilityZones= Oct 07 22:32:54 crc kubenswrapper[4871]: ++ EnableChassisAsGateway=true Oct 07 22:32:54 crc kubenswrapper[4871]: ++ PhysicalNetworks= Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNHostName= Oct 07 22:32:54 crc kubenswrapper[4871]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 07 22:32:54 crc kubenswrapper[4871]: ++ ovs_dir=/var/lib/openvswitch Oct 07 22:32:54 crc kubenswrapper[4871]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 07 22:32:54 crc kubenswrapper[4871]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 07 22:32:54 crc kubenswrapper[4871]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 07 22:32:54 crc kubenswrapper[4871]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 07 22:32:54 crc kubenswrapper[4871]: + sleep 0.5 Oct 07 22:32:54 crc kubenswrapper[4871]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 07 22:32:54 crc kubenswrapper[4871]: + cleanup_ovsdb_server_semaphore Oct 07 22:32:54 crc kubenswrapper[4871]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 07 22:32:54 crc kubenswrapper[4871]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 07 22:32:54 crc kubenswrapper[4871]: > Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.175183 4871 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 07 22:32:54 crc kubenswrapper[4871]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 07 22:32:54 crc kubenswrapper[4871]: + source /usr/local/bin/container-scripts/functions Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNBridge=br-int Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNRemote=tcp:localhost:6642 Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNEncapType=geneve Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNAvailabilityZones= Oct 07 22:32:54 crc kubenswrapper[4871]: ++ EnableChassisAsGateway=true Oct 07 22:32:54 crc kubenswrapper[4871]: ++ PhysicalNetworks= Oct 07 22:32:54 crc kubenswrapper[4871]: ++ OVNHostName= Oct 07 22:32:54 crc kubenswrapper[4871]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 07 22:32:54 crc kubenswrapper[4871]: ++ ovs_dir=/var/lib/openvswitch Oct 07 22:32:54 crc kubenswrapper[4871]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 07 22:32:54 crc kubenswrapper[4871]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 07 22:32:54 crc kubenswrapper[4871]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 07 22:32:54 crc kubenswrapper[4871]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 07 22:32:54 crc kubenswrapper[4871]: + sleep 0.5 Oct 07 22:32:54 crc kubenswrapper[4871]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 07 22:32:54 crc kubenswrapper[4871]: + cleanup_ovsdb_server_semaphore Oct 07 22:32:54 crc kubenswrapper[4871]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 07 22:32:54 crc kubenswrapper[4871]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 07 22:32:54 crc kubenswrapper[4871]: > pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" containerID="cri-o://2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.175239 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" containerID="cri-o://2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.194212 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-b2s89"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.211119 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-b2s89"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.219762 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f74a-account-create-6gc9r"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.229893 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f74a-account-create-6gc9r"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.236998 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderf74a-account-delete-mvrf4"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.242100 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "0a535bdf-4211-4f42-8137-ee421d6a7fae" (UID: "0a535bdf-4211-4f42-8137-ee421d6a7fae"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.252616 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7748587bfc-lxnz6"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.252938 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7748587bfc-lxnz6" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-httpd" containerID="cri-o://8348036e75afc1a3c2a675f95267987dea996bddef522f920d21ab3ccc437dfd" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.253071 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7748587bfc-lxnz6" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-server" containerID="cri-o://24dbb24a65b2912e92cb5795b73a40b47b50468929436754aa78a71953f92649" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.277177 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a79e-account-create-fwv68"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.286640 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.294684 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a79e-account-create-fwv68"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.322853 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-94p9x"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.325110 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-94p9x"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.330753 4871 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.330784 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.332227 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancea79e-account-delete-jv6lz"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.343660 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "0a535bdf-4211-4f42-8137-ee421d6a7fae" (UID: "0a535bdf-4211-4f42-8137-ee421d6a7fae"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.346080 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-9p2qr"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.359129 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-9p2qr"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.359927 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" containerID="cri-o://e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" gracePeriod=29 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.361785 4871 generic.go:334] "Generic (PLEG): container finished" podID="943e025c-54d8-4686-81ee-894259d7af56" containerID="0bf2fc0c8cbcc24b03cb64efbb163db4f05f24420ea760d649a0be66f64959ec" exitCode=137 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.379480 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vxlsj_ec7b131b-1100-4f58-bf1c-012ab142f03b/openstack-network-exporter/0.log" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.379644 4871 generic.go:334] "Generic (PLEG): container finished" podID="ec7b131b-1100-4f58-bf1c-012ab142f03b" containerID="0dce78a091797582beabb0faed2c660e6204b78527df1d3a95b8d0c04c1f1d92" exitCode=2 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.379843 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vxlsj" event={"ID":"ec7b131b-1100-4f58-bf1c-012ab142f03b","Type":"ContainerDied","Data":"0dce78a091797582beabb0faed2c660e6204b78527df1d3a95b8d0c04c1f1d92"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.388550 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vxlsj_ec7b131b-1100-4f58-bf1c-012ab142f03b/openstack-network-exporter/0.log" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.388628 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.395909 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7cc4f647bc-9lrsh"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.397230 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7cc4f647bc-9lrsh" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerName="neutron-api" containerID="cri-o://09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.397361 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7cc4f647bc-9lrsh" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerName="neutron-httpd" containerID="cri-o://48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.406126 4871 generic.go:334] "Generic (PLEG): container finished" podID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerID="893dcbed9540fca5394cb10bce40d68accf0e792ccb53a0bca71eeb8cf8337c3" exitCode=143 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.406337 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68dde0ec-a90a-4b25-b329-ec645d32671f","Type":"ContainerDied","Data":"893dcbed9540fca5394cb10bce40d68accf0e792ccb53a0bca71eeb8cf8337c3"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.412878 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-stzkd"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.420662 4871 generic.go:334] "Generic (PLEG): container finished" podID="9d299e74-bffe-4773-a33a-d615606ecc85" containerID="8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047" exitCode=143 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.420727 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9d299e74-bffe-4773-a33a-d615606ecc85","Type":"ContainerDied","Data":"8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047"} Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.428096 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6 is running failed: container process not found" containerID="d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.432558 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a535bdf-4211-4f42-8137-ee421d6a7fae-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.432643 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6 is running failed: container process not found" containerID="d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.432749 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-stzkd"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.444192 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-23c1-account-create-pgt7v"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.444273 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-23c1-account-create-pgt7v"] Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.448060 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6 is running failed: container process not found" containerID="d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.448136 4871 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="f9692529-979b-407e-9826-d93875feda52" containerName="ovsdbserver-nb" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.484591 4871 generic.go:334] "Generic (PLEG): container finished" podID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerID="0e1f568994d1e614f92908e5ccc4ef98ce4c080d84575da81deff6e86b817516" exitCode=143 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.484732 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23b48c5c-8fd6-4578-bc71-3f7b47860649","Type":"ContainerDied","Data":"0e1f568994d1e614f92908e5ccc4ef98ce4c080d84575da81deff6e86b817516"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.504652 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f9692529-979b-407e-9826-d93875feda52/ovsdbserver-nb/0.log" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.504717 4871 generic.go:334] "Generic (PLEG): container finished" podID="f9692529-979b-407e-9826-d93875feda52" containerID="e287d0f45913ff5f5dd60da81232d17cb6a0dd0bdcea624cee7faad1f4925c41" exitCode=2 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.504748 4871 generic.go:334] "Generic (PLEG): container finished" podID="f9692529-979b-407e-9826-d93875feda52" containerID="d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6" exitCode=143 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.505314 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f9692529-979b-407e-9826-d93875feda52","Type":"ContainerDied","Data":"e287d0f45913ff5f5dd60da81232d17cb6a0dd0bdcea624cee7faad1f4925c41"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.505355 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f9692529-979b-407e-9826-d93875feda52","Type":"ContainerDied","Data":"d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.513866 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-1599-account-create-ggqbn"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.543562 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ljmg\" (UniqueName: \"kubernetes.io/projected/ec7b131b-1100-4f58-bf1c-012ab142f03b-kube-api-access-5ljmg\") pod \"ec7b131b-1100-4f58-bf1c-012ab142f03b\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.543691 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovn-rundir\") pod \"ec7b131b-1100-4f58-bf1c-012ab142f03b\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.543855 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-metrics-certs-tls-certs\") pod \"ec7b131b-1100-4f58-bf1c-012ab142f03b\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.543893 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-combined-ca-bundle\") pod \"ec7b131b-1100-4f58-bf1c-012ab142f03b\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.543969 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7b131b-1100-4f58-bf1c-012ab142f03b-config\") pod \"ec7b131b-1100-4f58-bf1c-012ab142f03b\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.544111 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovs-rundir\") pod \"ec7b131b-1100-4f58-bf1c-012ab142f03b\" (UID: \"ec7b131b-1100-4f58-bf1c-012ab142f03b\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.545611 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "ec7b131b-1100-4f58-bf1c-012ab142f03b" (UID: "ec7b131b-1100-4f58-bf1c-012ab142f03b"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.547326 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-1599-account-create-ggqbn"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.553358 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "ec7b131b-1100-4f58-bf1c-012ab142f03b" (UID: "ec7b131b-1100-4f58-bf1c-012ab142f03b"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.554176 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec7b131b-1100-4f58-bf1c-012ab142f03b-config" (OuterVolumeSpecName: "config") pod "ec7b131b-1100-4f58-bf1c-012ab142f03b" (UID: "ec7b131b-1100-4f58-bf1c-012ab142f03b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.588050 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec7b131b-1100-4f58-bf1c-012ab142f03b-kube-api-access-5ljmg" (OuterVolumeSpecName: "kube-api-access-5ljmg") pod "ec7b131b-1100-4f58-bf1c-012ab142f03b" (UID: "ec7b131b-1100-4f58-bf1c-012ab142f03b"). InnerVolumeSpecName "kube-api-access-5ljmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.627922 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-7bj77"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630538 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630585 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630596 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630605 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630614 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630652 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630687 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630701 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630711 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630720 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630739 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630750 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630759 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630768 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630777 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630786 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630825 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630948 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630959 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.630969 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.653532 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="e1207599-6a0d-4019-83c0-8c0d6bed143b" containerName="galera" containerID="cri-o://91e6f4e8b318f1dbdfe9754e8b4775a9cd6b471b118d254f9da7064f27986523" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.648820 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec7b131b-1100-4f58-bf1c-012ab142f03b" (UID: "ec7b131b-1100-4f58-bf1c-012ab142f03b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.659123 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0a535bdf-4211-4f42-8137-ee421d6a7fae/ovsdbserver-sb/0.log" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.659298 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0a535bdf-4211-4f42-8137-ee421d6a7fae","Type":"ContainerDied","Data":"5320581b609b79547a868c9f4f5e9de107ecc6a2e52c7a7992ce5401278c8638"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.659332 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.659354 4871 scope.go:117] "RemoveContainer" containerID="d94cfb46e87419008029b8ee502b157dab2ee74b35e86d31ffb0102836fd1648" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.659952 4871 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.660756 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ljmg\" (UniqueName: \"kubernetes.io/projected/ec7b131b-1100-4f58-bf1c-012ab142f03b-kube-api-access-5ljmg\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.661382 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ec7b131b-1100-4f58-bf1c-012ab142f03b-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.661401 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.661413 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7b131b-1100-4f58-bf1c-012ab142f03b-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.663676 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f2b9-account-create-p2nlf"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.671551 4871 generic.go:334] "Generic (PLEG): container finished" podID="e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" containerID="36507cab7159abb01d536c353f20262fe1f5bacd4aef76bb370e449a8777f532" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.671630 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" event={"ID":"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e","Type":"ContainerDied","Data":"36507cab7159abb01d536c353f20262fe1f5bacd4aef76bb370e449a8777f532"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.677160 4871 generic.go:334] "Generic (PLEG): container finished" podID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" exitCode=0 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.677230 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jhh4r" event={"ID":"57b1b2c5-f33c-4af0-9707-168eae931b2d","Type":"ContainerDied","Data":"2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.679942 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f2b9-account-create-p2nlf"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.682338 4871 generic.go:334] "Generic (PLEG): container finished" podID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerID="303e52fe41cb9eb2c55a65b8e6fb2c26f6e89c632f0d230eb1f5ba71eb482fee" exitCode=143 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.682373 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d02631b4-f3a6-48d5-a43e-c90a48622ec2","Type":"ContainerDied","Data":"303e52fe41cb9eb2c55a65b8e6fb2c26f6e89c632f0d230eb1f5ba71eb482fee"} Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.708114 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementf2b9-account-delete-llr4x"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.708270 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-7bj77"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.751272 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "ec7b131b-1100-4f58-bf1c-012ab142f03b" (UID: "ec7b131b-1100-4f58-bf1c-012ab142f03b"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.760917 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.761866 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-7748587bfc-lxnz6" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.165:8080/healthcheck\": dial tcp 10.217.0.165:8080: connect: connection refused" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.761953 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-7748587bfc-lxnz6" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.165:8080/healthcheck\": dial tcp 10.217.0.165:8080: connect: connection refused" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.763473 4871 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec7b131b-1100-4f58-bf1c-012ab142f03b-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.790390 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-lrq4w"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.802147 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-d46f-account-create-7qnbt"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.817093 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-lrq4w"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.823334 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-d46f-account-create-7qnbt"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.830993 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapid46f-account-delete-8z75z"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.839627 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-s2nds"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.851745 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-719b-account-create-bx5fz"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.864822 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-s2nds"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.866874 4871 scope.go:117] "RemoveContainer" containerID="99e861992db1f9c7304f7a584cbddbe67197ed224df105828d0b35bd4e191a37" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.871696 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b245b\" (UniqueName: \"kubernetes.io/projected/943e025c-54d8-4686-81ee-894259d7af56-kube-api-access-b245b\") pod \"943e025c-54d8-4686-81ee-894259d7af56\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.871741 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-openstack-config-secret\") pod \"943e025c-54d8-4686-81ee-894259d7af56\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.871920 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-combined-ca-bundle\") pod \"943e025c-54d8-4686-81ee-894259d7af56\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.872063 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/943e025c-54d8-4686-81ee-894259d7af56-openstack-config\") pod \"943e025c-54d8-4686-81ee-894259d7af56\" (UID: \"943e025c-54d8-4686-81ee-894259d7af56\") " Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.876334 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-719b-account-create-bx5fz"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.891998 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.892365 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="c9e4de98-7d7d-43af-a76b-abf1774c620d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2ff1e6c6e2b4bf5a4e4fe91286071d2d9c47cf4d7640e8db226f47e4bb0157a4" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.898814 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/943e025c-54d8-4686-81ee-894259d7af56-kube-api-access-b245b" (OuterVolumeSpecName: "kube-api-access-b245b") pod "943e025c-54d8-4686-81ee-894259d7af56" (UID: "943e025c-54d8-4686-81ee-894259d7af56"). InnerVolumeSpecName "kube-api-access-b245b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.900527 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.918990 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6bc6797fd8-gflqx"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.919421 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" podUID="8724dd55-b509-4056-b324-531ec57c3a7e" containerName="barbican-keystone-listener-log" containerID="cri-o://d4cf3f49a66e3bfb1a2f0116e08e29835a98701bf050feff3371d723047d7b5c" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.919490 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" podUID="8724dd55-b509-4056-b324-531ec57c3a7e" containerName="barbican-keystone-listener" containerID="cri-o://2a05385f6943057613877aa85b019b47371856ca80b0e282c3f1b443a5093210" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.940035 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5544d45d85-dqp9x"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.940691 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5544d45d85-dqp9x" podUID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerName="barbican-worker-log" containerID="cri-o://26e20164e3525af55f84ab8ba8cc88a36fbcc0795cd362085e6b25b516d314d7" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.947494 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5544d45d85-dqp9x" podUID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerName="barbican-worker" containerID="cri-o://f1bb4a9d4f25de043a799d48cedbb3a8b9bba20dd97ad863253a098ddb8d0065" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.955338 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-677599d776-jjqd2"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.955692 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-677599d776-jjqd2" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api-log" containerID="cri-o://3ce17290b46def925cf8b9b69d4142ed8a8f644537f24587c21309531f91602e" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.955885 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-677599d776-jjqd2" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api" containerID="cri-o://e6f0d276a1846b36e45c6f4882a0d1c7b6023b4bc31e6f161d4fcc4d02c212f3" gracePeriod=30 Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.965514 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4vtwx"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.971779 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.974224 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4vtwx"] Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.977540 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b245b\" (UniqueName: \"kubernetes.io/projected/943e025c-54d8-4686-81ee-894259d7af56-kube-api-access-b245b\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.977615 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 07 22:32:54 crc kubenswrapper[4871]: E1007 22:32:54.977662 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data podName:1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb nodeName:}" failed. No retries permitted until 2025-10-07 22:32:56.977647343 +0000 UTC m=+1450.780345406 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data") pod "rabbitmq-cell1-server-0" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb") : configmap "rabbitmq-cell1-config-data" not found Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.979876 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f9692529-979b-407e-9826-d93875feda52/ovsdbserver-nb/0.log" Oct 07 22:32:54 crc kubenswrapper[4871]: I1007 22:32:54.979936 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.009546 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/943e025c-54d8-4686-81ee-894259d7af56-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "943e025c-54d8-4686-81ee-894259d7af56" (UID: "943e025c-54d8-4686-81ee-894259d7af56"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.016615 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a22909c-2279-4ea0-96b4-9abce934edc3" path="/var/lib/kubelet/pods/0a22909c-2279-4ea0-96b4-9abce934edc3/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.020765 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c59e7ea-43a5-4ae5-bd0d-36bdb761a042" path="/var/lib/kubelet/pods/0c59e7ea-43a5-4ae5-bd0d-36bdb761a042/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.022153 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10819804-8164-4a97-9fa0-fe2346c9f4f1" path="/var/lib/kubelet/pods/10819804-8164-4a97-9fa0-fe2346c9f4f1/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.022881 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1191dd97-3c15-4d04-b39a-ac4eec992a75" path="/var/lib/kubelet/pods/1191dd97-3c15-4d04-b39a-ac4eec992a75/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.023532 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="119e9627-9048-42c7-a8a1-5e0cf3167bce" path="/var/lib/kubelet/pods/119e9627-9048-42c7-a8a1-5e0cf3167bce/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.026845 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39ee3803-7a8c-4a85-b854-d72b5fe17dbb" path="/var/lib/kubelet/pods/39ee3803-7a8c-4a85-b854-d72b5fe17dbb/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.027501 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dea95fc-3d17-44db-bece-3bf7deaf387a" path="/var/lib/kubelet/pods/3dea95fc-3d17-44db-bece-3bf7deaf387a/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.028187 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7626d31e-6b79-44a4-b29d-0912f08ccbc7" path="/var/lib/kubelet/pods/7626d31e-6b79-44a4-b29d-0912f08ccbc7/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.029264 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88482765-93f6-4458-8511-1d932ebf2e16" path="/var/lib/kubelet/pods/88482765-93f6-4458-8511-1d932ebf2e16/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.029811 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dfe9a33-d304-430d-90dc-86a532e34080" path="/var/lib/kubelet/pods/8dfe9a33-d304-430d-90dc-86a532e34080/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.030394 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ee6fca7-2686-4de8-8b35-d00684e49fbc" path="/var/lib/kubelet/pods/8ee6fca7-2686-4de8-8b35-d00684e49fbc/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.031125 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4de1cd3-27f7-4a66-b5ec-89e5988b4d01" path="/var/lib/kubelet/pods/a4de1cd3-27f7-4a66-b5ec-89e5988b4d01/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.032356 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b13b0a19-d03a-4a5c-b77d-e95d97c534fd" path="/var/lib/kubelet/pods/b13b0a19-d03a-4a5c-b77d-e95d97c534fd/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.033068 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4f0df77-5bfb-4658-9349-b5089da279d8" path="/var/lib/kubelet/pods/b4f0df77-5bfb-4658-9349-b5089da279d8/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.033614 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b78d2643-0d33-49d1-9230-64c980990c7b" path="/var/lib/kubelet/pods/b78d2643-0d33-49d1-9230-64c980990c7b/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.034748 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd8851e2-da40-442d-a42b-d1a5cf56bbfa" path="/var/lib/kubelet/pods/bd8851e2-da40-442d-a42b-d1a5cf56bbfa/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.035278 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d" path="/var/lib/kubelet/pods/be6c62b2-d5e5-47fd-bff5-fc8a6c394c2d/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.035940 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c694d936-bdcc-4b7f-9bcd-f4552513dc0b" path="/var/lib/kubelet/pods/c694d936-bdcc-4b7f-9bcd-f4552513dc0b/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.036719 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce3ca1ea-9db5-4974-b974-ba4d25bf8e09" path="/var/lib/kubelet/pods/ce3ca1ea-9db5-4974-b974-ba4d25bf8e09/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.037779 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe40a70-0dba-4ccd-b540-9afbf5685d86" path="/var/lib/kubelet/pods/cfe40a70-0dba-4ccd-b540-9afbf5685d86/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.038363 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e124f5ed-d573-43b1-b93a-3d25367ef6b6" path="/var/lib/kubelet/pods/e124f5ed-d573-43b1-b93a-3d25367ef6b6/volumes" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.078477 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"f9692529-979b-407e-9826-d93875feda52\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.078572 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-metrics-certs-tls-certs\") pod \"f9692529-979b-407e-9826-d93875feda52\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.078633 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24kmx\" (UniqueName: \"kubernetes.io/projected/f9692529-979b-407e-9826-d93875feda52-kube-api-access-24kmx\") pod \"f9692529-979b-407e-9826-d93875feda52\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.078667 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-nb\") pod \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.078695 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-sb\") pod \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.078843 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-config\") pod \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.078872 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-svc\") pod \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.078916 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-scripts\") pod \"f9692529-979b-407e-9826-d93875feda52\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.078976 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9692529-979b-407e-9826-d93875feda52-ovsdb-rundir\") pod \"f9692529-979b-407e-9826-d93875feda52\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.079008 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-config\") pod \"f9692529-979b-407e-9826-d93875feda52\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.079030 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-combined-ca-bundle\") pod \"f9692529-979b-407e-9826-d93875feda52\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.079163 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd57k\" (UniqueName: \"kubernetes.io/projected/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-kube-api-access-hd57k\") pod \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.079207 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-ovsdbserver-nb-tls-certs\") pod \"f9692529-979b-407e-9826-d93875feda52\" (UID: \"f9692529-979b-407e-9826-d93875feda52\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.079242 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-swift-storage-0\") pod \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\" (UID: \"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e\") " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.079948 4871 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/943e025c-54d8-4686-81ee-894259d7af56-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.082573 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-config" (OuterVolumeSpecName: "config") pod "f9692529-979b-407e-9826-d93875feda52" (UID: "f9692529-979b-407e-9826-d93875feda52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.082961 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-scripts" (OuterVolumeSpecName: "scripts") pod "f9692529-979b-407e-9826-d93875feda52" (UID: "f9692529-979b-407e-9826-d93875feda52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.083125 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9692529-979b-407e-9826-d93875feda52-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "f9692529-979b-407e-9826-d93875feda52" (UID: "f9692529-979b-407e-9826-d93875feda52"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.117397 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9692529-979b-407e-9826-d93875feda52-kube-api-access-24kmx" (OuterVolumeSpecName: "kube-api-access-24kmx") pod "f9692529-979b-407e-9826-d93875feda52" (UID: "f9692529-979b-407e-9826-d93875feda52"). InnerVolumeSpecName "kube-api-access-24kmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.124150 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "f9692529-979b-407e-9826-d93875feda52" (UID: "f9692529-979b-407e-9826-d93875feda52"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.130220 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-kube-api-access-hd57k" (OuterVolumeSpecName: "kube-api-access-hd57k") pod "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" (UID: "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e"). InnerVolumeSpecName "kube-api-access-hd57k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.186031 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd57k\" (UniqueName: \"kubernetes.io/projected/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-kube-api-access-hd57k\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.186076 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.186086 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24kmx\" (UniqueName: \"kubernetes.io/projected/f9692529-979b-407e-9826-d93875feda52-kube-api-access-24kmx\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.186095 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.186104 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9692529-979b-407e-9826-d93875feda52-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.186113 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9692529-979b-407e-9826-d93875feda52-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.244572 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="dd77ac8e-36b6-4e16-869a-3ce6bff223ac" containerName="rabbitmq" containerID="cri-o://fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb" gracePeriod=604800 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.372969 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "943e025c-54d8-4686-81ee-894259d7af56" (UID: "943e025c-54d8-4686-81ee-894259d7af56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.393890 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.405389 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.426977 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "943e025c-54d8-4686-81ee-894259d7af56" (UID: "943e025c-54d8-4686-81ee-894259d7af56"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.450260 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9692529-979b-407e-9826-d93875feda52" (UID: "f9692529-979b-407e-9826-d93875feda52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.462659 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" (UID: "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.497775 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.497837 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.497849 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.497858 4871 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/943e025c-54d8-4686-81ee-894259d7af56-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.530514 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-config" (OuterVolumeSpecName: "config") pod "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" (UID: "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.565751 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" (UID: "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.585334 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" (UID: "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.604227 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.606708 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.606739 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.611382 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" (UID: "e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.619241 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "f9692529-979b-407e-9826-d93875feda52" (UID: "f9692529-979b-407e-9826-d93875feda52"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.647940 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "f9692529-979b-407e-9826-d93875feda52" (UID: "f9692529-979b-407e-9826-d93875feda52"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.720927 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.726126 4871 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.726183 4871 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9692529-979b-407e-9826-d93875feda52-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.743128 4871 generic.go:334] "Generic (PLEG): container finished" podID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerID="0b297eaadbd30609ad30865b36ac820cfdedfdd27b2cf2f2a38dea46f5c786d9" exitCode=143 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.750918 4871 generic.go:334] "Generic (PLEG): container finished" podID="361d6d85-f8c7-4643-95c4-29741f91a179" containerID="336725751a77fc6966bb612d7aba14caa7fe6292b10886ecbb69587c4cead039" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.777236 4871 generic.go:334] "Generic (PLEG): container finished" podID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerID="24dbb24a65b2912e92cb5795b73a40b47b50468929436754aa78a71953f92649" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.777272 4871 generic.go:334] "Generic (PLEG): container finished" podID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerID="8348036e75afc1a3c2a675f95267987dea996bddef522f920d21ab3ccc437dfd" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.781403 4871 generic.go:334] "Generic (PLEG): container finished" podID="e1207599-6a0d-4019-83c0-8c0d6bed143b" containerID="91e6f4e8b318f1dbdfe9754e8b4775a9cd6b471b118d254f9da7064f27986523" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.802221 4871 generic.go:334] "Generic (PLEG): container finished" podID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerID="48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.821874 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.821923 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.821933 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.821943 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.832301 4871 generic.go:334] "Generic (PLEG): container finished" podID="c750808e-de6d-442b-82ed-d3498b036d91" containerID="3ce17290b46def925cf8b9b69d4142ed8a8f644537f24587c21309531f91602e" exitCode=143 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.836028 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f9692529-979b-407e-9826-d93875feda52/ovsdbserver-nb/0.log" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.837040 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.839714 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.850841 4871 generic.go:334] "Generic (PLEG): container finished" podID="825fba74-f028-4445-b041-465ccbbad763" containerID="b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1" exitCode=0 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853462 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853497 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853519 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dbxmm"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853534 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f1ac984d-67d5-4bd1-b3c2-ef99976a8708","Type":"ContainerDied","Data":"0b297eaadbd30609ad30865b36ac820cfdedfdd27b2cf2f2a38dea46f5c786d9"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853556 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementf2b9-account-delete-llr4x" event={"ID":"361d6d85-f8c7-4643-95c4-29741f91a179","Type":"ContainerDied","Data":"336725751a77fc6966bb612d7aba14caa7fe6292b10886ecbb69587c4cead039"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853574 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dbxmm"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853591 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853605 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementf2b9-account-delete-llr4x"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853617 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementf2b9-account-delete-llr4x" event={"ID":"361d6d85-f8c7-4643-95c4-29741f91a179","Type":"ContainerStarted","Data":"abc48b0a08fe0a95b0be6d777e6f82b72eb974a6cb4df9a8bc43ab2a4798f7b5"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853628 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7748587bfc-lxnz6" event={"ID":"645a6027-7304-4fdc-aaf4-ffd433bec275","Type":"ContainerDied","Data":"24dbb24a65b2912e92cb5795b73a40b47b50468929436754aa78a71953f92649"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853643 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7748587bfc-lxnz6" event={"ID":"645a6027-7304-4fdc-aaf4-ffd433bec275","Type":"ContainerDied","Data":"8348036e75afc1a3c2a675f95267987dea996bddef522f920d21ab3ccc437dfd"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853655 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7748587bfc-lxnz6" event={"ID":"645a6027-7304-4fdc-aaf4-ffd433bec275","Type":"ContainerDied","Data":"ac679d9e737a4894873c1cb1029ff1234f15dd5986c1ace6c6d5154d10fd88bf"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853666 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac679d9e737a4894873c1cb1029ff1234f15dd5986c1ace6c6d5154d10fd88bf" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853679 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e1207599-6a0d-4019-83c0-8c0d6bed143b","Type":"ContainerDied","Data":"91e6f4e8b318f1dbdfe9754e8b4775a9cd6b471b118d254f9da7064f27986523"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853693 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853707 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancea79e-account-delete-jv6lz"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853717 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapid46f-account-delete-8z75z"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853727 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cc4f647bc-9lrsh" event={"ID":"7116e0fa-30d9-4a8d-9a11-0d57908f268f","Type":"ContainerDied","Data":"48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853738 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderf74a-account-delete-mvrf4"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853750 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853761 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853776 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853788 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.853825 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.854451 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="1e61b0b6-83cc-48b0-9f3d-b9b607faa830" containerName="nova-cell0-conductor-conductor" containerID="cri-o://655d7d7e1b7fb65057965cba90062d42b1f499d86be0b09af648ef03eef592b5" gracePeriod=30 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.854687 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.854718 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677599d776-jjqd2" event={"ID":"c750808e-de6d-442b-82ed-d3498b036d91","Type":"ContainerDied","Data":"3ce17290b46def925cf8b9b69d4142ed8a8f644537f24587c21309531f91602e"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.854746 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f9692529-979b-407e-9826-d93875feda52","Type":"ContainerDied","Data":"1f2bc1440a98fedcc969594e5cf11b2d39214fdc0054993e38321b50dd3f4099"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.854771 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapid46f-account-delete-8z75z" event={"ID":"ca98e71f-5147-4bb7-bd76-e4bb11cfcf42","Type":"ContainerStarted","Data":"58de1ef62139407f0ad019c635b6f4e1d1c507b22628865fb44d4e6cb0dbf48f"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.854808 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"825fba74-f028-4445-b041-465ccbbad763","Type":"ContainerDied","Data":"b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.854845 4871 scope.go:117] "RemoveContainer" containerID="e287d0f45913ff5f5dd60da81232d17cb6a0dd0bdcea624cee7faad1f4925c41" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.857171 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderf74a-account-delete-mvrf4" event={"ID":"1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10","Type":"ContainerStarted","Data":"e33c1e5fc4fb38799f00d2cd45f268d07a9d843029bb931d15df8f5ed4a7cf65"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.857284 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="fd68bd87-c8ef-4214-a9b2-299efdcb956e" containerName="nova-cell1-conductor-conductor" containerID="cri-o://f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63" gracePeriod=30 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.859036 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ab7d1def-9358-44f0-9245-e333fb60bac4" containerName="nova-scheduler-scheduler" containerID="cri-o://d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" gracePeriod=30 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.875954 4871 generic.go:334] "Generic (PLEG): container finished" podID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerID="26e20164e3525af55f84ab8ba8cc88a36fbcc0795cd362085e6b25b516d314d7" exitCode=143 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.876094 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5544d45d85-dqp9x" event={"ID":"f9973055-9eec-4599-9d2b-ccef87f6a481","Type":"ContainerDied","Data":"26e20164e3525af55f84ab8ba8cc88a36fbcc0795cd362085e6b25b516d314d7"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.887318 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" event={"ID":"e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e","Type":"ContainerDied","Data":"b8b4f077f6ba2b4943c091859f32d76cb65df0d33c978f83c2370116ac851bdb"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.887453 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5967cc9597-phtsg" Oct 07 22:32:55 crc kubenswrapper[4871]: E1007 22:32:55.913386 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.920851 4871 generic.go:334] "Generic (PLEG): container finished" podID="878958ac-64b8-434b-88ee-f23d3c486765" containerID="ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc" exitCode=143 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.920924 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7567f46b5d-kb4h9" event={"ID":"878958ac-64b8-434b-88ee-f23d3c486765","Type":"ContainerDied","Data":"ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.928056 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.933719 4871 generic.go:334] "Generic (PLEG): container finished" podID="8724dd55-b509-4056-b324-531ec57c3a7e" containerID="d4cf3f49a66e3bfb1a2f0116e08e29835a98701bf050feff3371d723047d7b5c" exitCode=143 Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.934055 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" event={"ID":"8724dd55-b509-4056-b324-531ec57c3a7e","Type":"ContainerDied","Data":"d4cf3f49a66e3bfb1a2f0116e08e29835a98701bf050feff3371d723047d7b5c"} Oct 07 22:32:55 crc kubenswrapper[4871]: E1007 22:32:55.971098 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.972500 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vxlsj_ec7b131b-1100-4f58-bf1c-012ab142f03b/openstack-network-exporter/0.log" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.972740 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vxlsj" event={"ID":"ec7b131b-1100-4f58-bf1c-012ab142f03b","Type":"ContainerDied","Data":"1dee67987aed975e2f4f7e026b61a444e730059988839ac8f1623131ce02236f"} Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.972932 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vxlsj" Oct 07 22:32:55 crc kubenswrapper[4871]: E1007 22:32:55.974294 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 22:32:55 crc kubenswrapper[4871]: E1007 22:32:55.974371 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ab7d1def-9358-44f0-9245-e333fb60bac4" containerName="nova-scheduler-scheduler" Oct 07 22:32:55 crc kubenswrapper[4871]: I1007 22:32:55.980336 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" containerName="rabbitmq" containerID="cri-o://798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d" gracePeriod=604800 Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:55.997976 4871 scope.go:117] "RemoveContainer" containerID="d1e814280c2541d014312d973be4504cdcf1952d40c4ffdef8ca6d8fb7a619b6" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.020552 4871 generic.go:334] "Generic (PLEG): container finished" podID="b17ba7e1-deb0-4a31-8874-ee1e6686a3f8" containerID="b1ba6099d51e0bd9f15eb3e0c286b7e8227ca05e044b331b744976e0d99a05b9" exitCode=0 Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.020652 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancea79e-account-delete-jv6lz" event={"ID":"b17ba7e1-deb0-4a31-8874-ee1e6686a3f8","Type":"ContainerDied","Data":"b1ba6099d51e0bd9f15eb3e0c286b7e8227ca05e044b331b744976e0d99a05b9"} Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.020709 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancea79e-account-delete-jv6lz" event={"ID":"b17ba7e1-deb0-4a31-8874-ee1e6686a3f8","Type":"ContainerStarted","Data":"0e35dde0ab66114c6402e22ac0f0bef2f78194e908aec98110d5f9c302ac45e6"} Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.041867 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-public-tls-certs\") pod \"645a6027-7304-4fdc-aaf4-ffd433bec275\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.042067 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-log-httpd\") pod \"645a6027-7304-4fdc-aaf4-ffd433bec275\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.042202 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-internal-tls-certs\") pod \"645a6027-7304-4fdc-aaf4-ffd433bec275\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.042248 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-etc-swift\") pod \"645a6027-7304-4fdc-aaf4-ffd433bec275\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.042301 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-combined-ca-bundle\") pod \"645a6027-7304-4fdc-aaf4-ffd433bec275\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.042370 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-run-httpd\") pod \"645a6027-7304-4fdc-aaf4-ffd433bec275\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.042443 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-config-data\") pod \"645a6027-7304-4fdc-aaf4-ffd433bec275\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.042506 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rxdn\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-kube-api-access-7rxdn\") pod \"645a6027-7304-4fdc-aaf4-ffd433bec275\" (UID: \"645a6027-7304-4fdc-aaf4-ffd433bec275\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.043075 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "645a6027-7304-4fdc-aaf4-ffd433bec275" (UID: "645a6027-7304-4fdc-aaf4-ffd433bec275"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.047261 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "645a6027-7304-4fdc-aaf4-ffd433bec275" (UID: "645a6027-7304-4fdc-aaf4-ffd433bec275"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.061190 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.061227 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/645a6027-7304-4fdc-aaf4-ffd433bec275-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.068432 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.080906 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.091758 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-phtsg"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.100607 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-phtsg"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.117096 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-kube-api-access-7rxdn" (OuterVolumeSpecName: "kube-api-access-7rxdn") pod "645a6027-7304-4fdc-aaf4-ffd433bec275" (UID: "645a6027-7304-4fdc-aaf4-ffd433bec275"). InnerVolumeSpecName "kube-api-access-7rxdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.130202 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.130726 4871 scope.go:117] "RemoveContainer" containerID="0bf2fc0c8cbcc24b03cb64efbb163db4f05f24420ea760d649a0be66f64959ec" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.160338 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-vxlsj"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.163649 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-kolla-config\") pod \"e1207599-6a0d-4019-83c0-8c0d6bed143b\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.163707 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-generated\") pod \"e1207599-6a0d-4019-83c0-8c0d6bed143b\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.163773 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-combined-ca-bundle\") pod \"e1207599-6a0d-4019-83c0-8c0d6bed143b\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.164044 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-default\") pod \"e1207599-6a0d-4019-83c0-8c0d6bed143b\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.164105 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"e1207599-6a0d-4019-83c0-8c0d6bed143b\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.164180 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxzl4\" (UniqueName: \"kubernetes.io/projected/e1207599-6a0d-4019-83c0-8c0d6bed143b-kube-api-access-xxzl4\") pod \"e1207599-6a0d-4019-83c0-8c0d6bed143b\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.164250 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-secrets\") pod \"e1207599-6a0d-4019-83c0-8c0d6bed143b\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.164276 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-galera-tls-certs\") pod \"e1207599-6a0d-4019-83c0-8c0d6bed143b\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.164337 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-operator-scripts\") pod \"e1207599-6a0d-4019-83c0-8c0d6bed143b\" (UID: \"e1207599-6a0d-4019-83c0-8c0d6bed143b\") " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.164965 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rxdn\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-kube-api-access-7rxdn\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.166261 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e1207599-6a0d-4019-83c0-8c0d6bed143b" (UID: "e1207599-6a0d-4019-83c0-8c0d6bed143b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.167100 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "e1207599-6a0d-4019-83c0-8c0d6bed143b" (UID: "e1207599-6a0d-4019-83c0-8c0d6bed143b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.167544 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "e1207599-6a0d-4019-83c0-8c0d6bed143b" (UID: "e1207599-6a0d-4019-83c0-8c0d6bed143b"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.169551 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.169608 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data podName:dd77ac8e-36b6-4e16-869a-3ce6bff223ac nodeName:}" failed. No retries permitted until 2025-10-07 22:33:00.169590767 +0000 UTC m=+1453.972288830 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data") pod "rabbitmq-server-0" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac") : configmap "rabbitmq-config-data" not found Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.170785 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "e1207599-6a0d-4019-83c0-8c0d6bed143b" (UID: "e1207599-6a0d-4019-83c0-8c0d6bed143b"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.182331 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "645a6027-7304-4fdc-aaf4-ffd433bec275" (UID: "645a6027-7304-4fdc-aaf4-ffd433bec275"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.182431 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-vxlsj"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.204430 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-secrets" (OuterVolumeSpecName: "secrets") pod "e1207599-6a0d-4019-83c0-8c0d6bed143b" (UID: "e1207599-6a0d-4019-83c0-8c0d6bed143b"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.245832 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.246168 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="ceilometer-central-agent" containerID="cri-o://b2140ff400665da608cd7a343da108840518d021eb0a870d34ed4a85d295f742" gracePeriod=30 Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.246687 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="proxy-httpd" containerID="cri-o://457a86497864d11e171b8cdea112cccb3d4066b22a76ceab718229ee7ea9a56c" gracePeriod=30 Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.246734 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="sg-core" containerID="cri-o://b1bc98f764cc4f92e14878e9fcdb1b85ceb45251562f19cc562511bd4c01c7c5" gracePeriod=30 Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.246762 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="ceilometer-notification-agent" containerID="cri-o://f02b2caa39b480dc045d7e7262686751f44a5f1546d461810390f4aa68b67b32" gracePeriod=30 Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.269270 4871 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.269301 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.269311 4871 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/645a6027-7304-4fdc-aaf4-ffd433bec275-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.269319 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.269328 4871 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.269336 4871 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1207599-6a0d-4019-83c0-8c0d6bed143b-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.272087 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1207599-6a0d-4019-83c0-8c0d6bed143b-kube-api-access-xxzl4" (OuterVolumeSpecName: "kube-api-access-xxzl4") pod "e1207599-6a0d-4019-83c0-8c0d6bed143b" (UID: "e1207599-6a0d-4019-83c0-8c0d6bed143b"). InnerVolumeSpecName "kube-api-access-xxzl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.329128 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.329377 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="829bd903-9e3e-4378-9c56-6a6375d26422" containerName="kube-state-metrics" containerID="cri-o://1b18488a0daf1c9ccb071f5a1f222406f2336c9c403fbfe6cbfc92c70b53edf6" gracePeriod=30 Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.361603 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "mysql-db") pod "e1207599-6a0d-4019-83c0-8c0d6bed143b" (UID: "e1207599-6a0d-4019-83c0-8c0d6bed143b"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.381955 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.381997 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxzl4\" (UniqueName: \"kubernetes.io/projected/e1207599-6a0d-4019-83c0-8c0d6bed143b-kube-api-access-xxzl4\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.627280 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.695679 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.732647 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.747941 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="f602e068-0d3e-4f32-91ca-3ec36c965a7c" containerName="memcached" containerID="cri-o://cd6faea6c3b696458cd8465298585cf7714badd7a13ee982186f92932560d733" gracePeriod=30 Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.807686 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone6eb9-account-delete-7xq7t"] Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808607 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" containerName="init" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808622 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" containerName="init" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808641 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-server" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808648 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-server" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808658 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-httpd" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808664 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-httpd" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808680 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1207599-6a0d-4019-83c0-8c0d6bed143b" containerName="galera" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808687 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1207599-6a0d-4019-83c0-8c0d6bed143b" containerName="galera" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808709 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerName="ovsdbserver-sb" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808716 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerName="ovsdbserver-sb" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808728 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerName="openstack-network-exporter" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808734 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerName="openstack-network-exporter" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808754 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1207599-6a0d-4019-83c0-8c0d6bed143b" containerName="mysql-bootstrap" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808760 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1207599-6a0d-4019-83c0-8c0d6bed143b" containerName="mysql-bootstrap" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808776 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec7b131b-1100-4f58-bf1c-012ab142f03b" containerName="openstack-network-exporter" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808781 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec7b131b-1100-4f58-bf1c-012ab142f03b" containerName="openstack-network-exporter" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808813 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9692529-979b-407e-9826-d93875feda52" containerName="openstack-network-exporter" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808821 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9692529-979b-407e-9826-d93875feda52" containerName="openstack-network-exporter" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808829 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9692529-979b-407e-9826-d93875feda52" containerName="ovsdbserver-nb" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808839 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9692529-979b-407e-9826-d93875feda52" containerName="ovsdbserver-nb" Oct 07 22:32:56 crc kubenswrapper[4871]: E1007 22:32:56.808853 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" containerName="dnsmasq-dns" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.808861 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" containerName="dnsmasq-dns" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.816722 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9692529-979b-407e-9826-d93875feda52" containerName="openstack-network-exporter" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.816767 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerName="openstack-network-exporter" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.816784 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-httpd" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.816823 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a535bdf-4211-4f42-8137-ee421d6a7fae" containerName="ovsdbserver-sb" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.816835 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec7b131b-1100-4f58-bf1c-012ab142f03b" containerName="openstack-network-exporter" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.816848 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9692529-979b-407e-9826-d93875feda52" containerName="ovsdbserver-nb" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.816866 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1207599-6a0d-4019-83c0-8c0d6bed143b" containerName="galera" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.816884 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" containerName="proxy-server" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.816900 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" containerName="dnsmasq-dns" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.818078 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone6eb9-account-delete-7xq7t" Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.850539 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-7zfdn"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.877889 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xt474"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.946141 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-7zfdn"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.955668 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone6eb9-account-delete-7xq7t"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.965769 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xt474"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.972898 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-cd55cdc75-kj6jh"] Oct 07 22:32:56 crc kubenswrapper[4871]: I1007 22:32:56.973304 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-cd55cdc75-kj6jh" podUID="6af0bcb3-3803-4174-af06-c6841610c62a" containerName="keystone-api" containerID="cri-o://1987ababc57bf8086a0c131045faa2ea781c10edf009a458861520b942f65314" gracePeriod=30 Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.019004 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stmjr\" (UniqueName: \"kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr\") pod \"keystone6eb9-account-delete-7xq7t\" (UID: \"27550e09-71f4-4ed9-96c1-fd40f3f10cfb\") " pod="openstack/keystone6eb9-account-delete-7xq7t" Oct 07 22:32:57 crc kubenswrapper[4871]: E1007 22:32:57.019738 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 07 22:32:57 crc kubenswrapper[4871]: E1007 22:32:57.019829 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data podName:1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb nodeName:}" failed. No retries permitted until 2025-10-07 22:33:01.019807674 +0000 UTC m=+1454.822505747 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data") pod "rabbitmq-cell1-server-0" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb") : configmap "rabbitmq-cell1-config-data" not found Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.032415 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="034a9072-67f8-45cb-896a-b238766f2f16" path="/var/lib/kubelet/pods/034a9072-67f8-45cb-896a-b238766f2f16/volumes" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.033684 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a535bdf-4211-4f42-8137-ee421d6a7fae" path="/var/lib/kubelet/pods/0a535bdf-4211-4f42-8137-ee421d6a7fae/volumes" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.034556 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22863eb4-bd2f-42d6-a60c-5489bf5c1c81" path="/var/lib/kubelet/pods/22863eb4-bd2f-42d6-a60c-5489bf5c1c81/volumes" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.041727 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="943e025c-54d8-4686-81ee-894259d7af56" path="/var/lib/kubelet/pods/943e025c-54d8-4686-81ee-894259d7af56/volumes" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.045071 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e" path="/var/lib/kubelet/pods/e5d55d4b-6e19-46ed-9ca8-04ddab7c8b2e/volumes" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.046745 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec7b131b-1100-4f58-bf1c-012ab142f03b" path="/var/lib/kubelet/pods/ec7b131b-1100-4f58-bf1c-012ab142f03b/volumes" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.058495 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f23add1a-b7b7-4670-b510-985a7c4c221d" path="/var/lib/kubelet/pods/f23add1a-b7b7-4670-b510-985a7c4c221d/volumes" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.060868 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9692529-979b-407e-9826-d93875feda52" path="/var/lib/kubelet/pods/f9692529-979b-407e-9826-d93875feda52/volumes" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.099632 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.164:8776/healthcheck\": read tcp 10.217.0.2:48496->10.217.0.164:8776: read: connection reset by peer" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.138527 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stmjr\" (UniqueName: \"kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr\") pod \"keystone6eb9-account-delete-7xq7t\" (UID: \"27550e09-71f4-4ed9-96c1-fd40f3f10cfb\") " pod="openstack/keystone6eb9-account-delete-7xq7t" Oct 07 22:32:57 crc kubenswrapper[4871]: E1007 22:32:57.146821 4871 projected.go:194] Error preparing data for projected volume kube-api-access-stmjr for pod openstack/keystone6eb9-account-delete-7xq7t: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 07 22:32:57 crc kubenswrapper[4871]: E1007 22:32:57.146886 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr podName:27550e09-71f4-4ed9-96c1-fd40f3f10cfb nodeName:}" failed. No retries permitted until 2025-10-07 22:32:57.646865543 +0000 UTC m=+1451.449563616 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-stmjr" (UniqueName: "kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr") pod "keystone6eb9-account-delete-7xq7t" (UID: "27550e09-71f4-4ed9-96c1-fd40f3f10cfb") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.164421 4871 generic.go:334] "Generic (PLEG): container finished" podID="1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10" containerID="f988c8bf25a955bb983d8bdd6ca560c46444fc402fe7eefe2c5fdbd7ae80b4c5" exitCode=0 Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.183509 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "645a6027-7304-4fdc-aaf4-ffd433bec275" (UID: "645a6027-7304-4fdc-aaf4-ffd433bec275"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.200390 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "645a6027-7304-4fdc-aaf4-ffd433bec275" (UID: "645a6027-7304-4fdc-aaf4-ffd433bec275"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.202754 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "645a6027-7304-4fdc-aaf4-ffd433bec275" (UID: "645a6027-7304-4fdc-aaf4-ffd433bec275"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.225452 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1207599-6a0d-4019-83c0-8c0d6bed143b" (UID: "e1207599-6a0d-4019-83c0-8c0d6bed143b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.243045 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.243099 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.243109 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.243118 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.247122 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-config-data" (OuterVolumeSpecName: "config-data") pod "645a6027-7304-4fdc-aaf4-ffd433bec275" (UID: "645a6027-7304-4fdc-aaf4-ffd433bec275"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.269032 4871 generic.go:334] "Generic (PLEG): container finished" podID="829bd903-9e3e-4378-9c56-6a6375d26422" containerID="1b18488a0daf1c9ccb071f5a1f222406f2336c9c403fbfe6cbfc92c70b53edf6" exitCode=2 Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.294114 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "e1207599-6a0d-4019-83c0-8c0d6bed143b" (UID: "e1207599-6a0d-4019-83c0-8c0d6bed143b"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.345224 4871 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1207599-6a0d-4019-83c0-8c0d6bed143b-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.345257 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645a6027-7304-4fdc-aaf4-ffd433bec275-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.348823 4871 generic.go:334] "Generic (PLEG): container finished" podID="e71edefc-c645-4e74-8077-42df9e751ce0" containerID="b1bc98f764cc4f92e14878e9fcdb1b85ceb45251562f19cc562511bd4c01c7c5" exitCode=2 Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.359484 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.366543 4871 generic.go:334] "Generic (PLEG): container finished" podID="c9e4de98-7d7d-43af-a76b-abf1774c620d" containerID="2ff1e6c6e2b4bf5a4e4fe91286071d2d9c47cf4d7640e8db226f47e4bb0157a4" exitCode=0 Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.366695 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.471537 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": read tcp 10.217.0.2:40306->10.217.0.205:8775: read: connection reset by peer" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.472058 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": read tcp 10.217.0.2:40300->10.217.0.205:8775: read: connection reset by peer" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.653573 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stmjr\" (UniqueName: \"kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr\") pod \"keystone6eb9-account-delete-7xq7t\" (UID: \"27550e09-71f4-4ed9-96c1-fd40f3f10cfb\") " pod="openstack/keystone6eb9-account-delete-7xq7t" Oct 07 22:32:57 crc kubenswrapper[4871]: E1007 22:32:57.658603 4871 projected.go:194] Error preparing data for projected volume kube-api-access-stmjr for pod openstack/keystone6eb9-account-delete-7xq7t: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 07 22:32:57 crc kubenswrapper[4871]: E1007 22:32:57.658682 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr podName:27550e09-71f4-4ed9-96c1-fd40f3f10cfb nodeName:}" failed. No retries permitted until 2025-10-07 22:32:58.658657691 +0000 UTC m=+1452.461355764 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-stmjr" (UniqueName: "kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr") pod "keystone6eb9-account-delete-7xq7t" (UID: "27550e09-71f4-4ed9-96c1-fd40f3f10cfb") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.791311 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.200:3000/\": dial tcp 10.217.0.200:3000: connect: connection refused" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.826743 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.826815 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderf74a-account-delete-mvrf4" event={"ID":"1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10","Type":"ContainerDied","Data":"f988c8bf25a955bb983d8bdd6ca560c46444fc402fe7eefe2c5fdbd7ae80b4c5"} Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.826897 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-wqf47"] Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.826914 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementf2b9-account-delete-llr4x" event={"ID":"361d6d85-f8c7-4643-95c4-29741f91a179","Type":"ContainerDied","Data":"abc48b0a08fe0a95b0be6d777e6f82b72eb974a6cb4df9a8bc43ab2a4798f7b5"} Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.826931 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abc48b0a08fe0a95b0be6d777e6f82b72eb974a6cb4df9a8bc43ab2a4798f7b5" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.826946 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-wqf47"] Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.826964 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6eb9-account-create-ngxgg"] Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.826976 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone6eb9-account-delete-7xq7t"] Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.826993 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"829bd903-9e3e-4378-9c56-6a6375d26422","Type":"ContainerDied","Data":"1b18488a0daf1c9ccb071f5a1f222406f2336c9c403fbfe6cbfc92c70b53edf6"} Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.827010 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancea79e-account-delete-jv6lz" event={"ID":"b17ba7e1-deb0-4a31-8874-ee1e6686a3f8","Type":"ContainerDied","Data":"0e35dde0ab66114c6402e22ac0f0bef2f78194e908aec98110d5f9c302ac45e6"} Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.827023 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e35dde0ab66114c6402e22ac0f0bef2f78194e908aec98110d5f9c302ac45e6" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.827034 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6eb9-account-create-ngxgg"] Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.827048 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerDied","Data":"b1bc98f764cc4f92e14878e9fcdb1b85ceb45251562f19cc562511bd4c01c7c5"} Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.827064 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e1207599-6a0d-4019-83c0-8c0d6bed143b","Type":"ContainerDied","Data":"b152470c4739f052ea68c8b5f32ead55450b0adba08d66c2d110efa3590e6304"} Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.827083 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c9e4de98-7d7d-43af-a76b-abf1774c620d","Type":"ContainerDied","Data":"2ff1e6c6e2b4bf5a4e4fe91286071d2d9c47cf4d7640e8db226f47e4bb0157a4"} Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.827097 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c9e4de98-7d7d-43af-a76b-abf1774c620d","Type":"ContainerDied","Data":"f3956bb5766dd205fa72ec949c6e4fca7a5b0c2ca288def1748431f4ffa0258d"} Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.827107 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3956bb5766dd205fa72ec949c6e4fca7a5b0c2ca288def1748431f4ffa0258d" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.847047 4871 scope.go:117] "RemoveContainer" containerID="36507cab7159abb01d536c353f20262fe1f5bacd4aef76bb370e449a8777f532" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.917934 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementf2b9-account-delete-llr4x" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.957497 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9wcx\" (UniqueName: \"kubernetes.io/projected/361d6d85-f8c7-4643-95c4-29741f91a179-kube-api-access-r9wcx\") pod \"361d6d85-f8c7-4643-95c4-29741f91a179\" (UID: \"361d6d85-f8c7-4643-95c4-29741f91a179\") " Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.965194 4871 scope.go:117] "RemoveContainer" containerID="33f7ea0548fe776e435826efc31be59ac4fe01983421306e3a6bb36a95b9f477" Oct 07 22:32:57 crc kubenswrapper[4871]: E1007 22:32:57.972861 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-stmjr], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone6eb9-account-delete-7xq7t" podUID="27550e09-71f4-4ed9-96c1-fd40f3f10cfb" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.988324 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancea79e-account-delete-jv6lz" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.991226 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/361d6d85-f8c7-4643-95c4-29741f91a179-kube-api-access-r9wcx" (OuterVolumeSpecName: "kube-api-access-r9wcx") pod "361d6d85-f8c7-4643-95c4-29741f91a179" (UID: "361d6d85-f8c7-4643-95c4-29741f91a179"). InnerVolumeSpecName "kube-api-access-r9wcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:57 crc kubenswrapper[4871]: I1007 22:32:57.997012 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.001903 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.007056 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.015707 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderf74a-account-delete-mvrf4" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.017219 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.043074 4871 scope.go:117] "RemoveContainer" containerID="0dce78a091797582beabb0faed2c660e6204b78527df1d3a95b8d0c04c1f1d92" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.044618 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.061185 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9wcx\" (UniqueName: \"kubernetes.io/projected/361d6d85-f8c7-4643-95c4-29741f91a179-kube-api-access-r9wcx\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.093017 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.141094 4871 scope.go:117] "RemoveContainer" containerID="91e6f4e8b318f1dbdfe9754e8b4775a9cd6b471b118d254f9da7064f27986523" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163272 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fmm8\" (UniqueName: \"kubernetes.io/projected/c9e4de98-7d7d-43af-a76b-abf1774c620d-kube-api-access-6fmm8\") pod \"c9e4de98-7d7d-43af-a76b-abf1774c620d\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163317 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-nova-novncproxy-tls-certs\") pod \"c9e4de98-7d7d-43af-a76b-abf1774c620d\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163342 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d299e74-bffe-4773-a33a-d615606ecc85-logs\") pod \"9d299e74-bffe-4773-a33a-d615606ecc85\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163357 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-scripts\") pod \"9d299e74-bffe-4773-a33a-d615606ecc85\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163387 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-vencrypt-tls-certs\") pod \"c9e4de98-7d7d-43af-a76b-abf1774c620d\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163435 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7748t\" (UniqueName: \"kubernetes.io/projected/9d299e74-bffe-4773-a33a-d615606ecc85-kube-api-access-7748t\") pod \"9d299e74-bffe-4773-a33a-d615606ecc85\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163455 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data\") pod \"9d299e74-bffe-4773-a33a-d615606ecc85\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163480 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-combined-ca-bundle\") pod \"829bd903-9e3e-4378-9c56-6a6375d26422\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163496 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data-custom\") pod \"9d299e74-bffe-4773-a33a-d615606ecc85\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163563 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkrz5\" (UniqueName: \"kubernetes.io/projected/1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10-kube-api-access-nkrz5\") pod \"1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10\" (UID: \"1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163585 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d299e74-bffe-4773-a33a-d615606ecc85-etc-machine-id\") pod \"9d299e74-bffe-4773-a33a-d615606ecc85\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163604 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-certs\") pod \"829bd903-9e3e-4378-9c56-6a6375d26422\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163625 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-public-tls-certs\") pod \"9d299e74-bffe-4773-a33a-d615606ecc85\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163642 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz4t2\" (UniqueName: \"kubernetes.io/projected/829bd903-9e3e-4378-9c56-6a6375d26422-kube-api-access-tz4t2\") pod \"829bd903-9e3e-4378-9c56-6a6375d26422\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163702 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-combined-ca-bundle\") pod \"9d299e74-bffe-4773-a33a-d615606ecc85\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163743 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-internal-tls-certs\") pod \"9d299e74-bffe-4773-a33a-d615606ecc85\" (UID: \"9d299e74-bffe-4773-a33a-d615606ecc85\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163764 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-config-data\") pod \"c9e4de98-7d7d-43af-a76b-abf1774c620d\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163820 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-combined-ca-bundle\") pod \"c9e4de98-7d7d-43af-a76b-abf1774c620d\" (UID: \"c9e4de98-7d7d-43af-a76b-abf1774c620d\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163837 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-config\") pod \"829bd903-9e3e-4378-9c56-6a6375d26422\" (UID: \"829bd903-9e3e-4378-9c56-6a6375d26422\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.163880 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdjq5\" (UniqueName: \"kubernetes.io/projected/b17ba7e1-deb0-4a31-8874-ee1e6686a3f8-kube-api-access-wdjq5\") pod \"b17ba7e1-deb0-4a31-8874-ee1e6686a3f8\" (UID: \"b17ba7e1-deb0-4a31-8874-ee1e6686a3f8\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.169737 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d299e74-bffe-4773-a33a-d615606ecc85-logs" (OuterVolumeSpecName: "logs") pod "9d299e74-bffe-4773-a33a-d615606ecc85" (UID: "9d299e74-bffe-4773-a33a-d615606ecc85"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.176335 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d299e74-bffe-4773-a33a-d615606ecc85-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9d299e74-bffe-4773-a33a-d615606ecc85" (UID: "9d299e74-bffe-4773-a33a-d615606ecc85"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.183931 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9d299e74-bffe-4773-a33a-d615606ecc85" (UID: "9d299e74-bffe-4773-a33a-d615606ecc85"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.184250 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-scripts" (OuterVolumeSpecName: "scripts") pod "9d299e74-bffe-4773-a33a-d615606ecc85" (UID: "9d299e74-bffe-4773-a33a-d615606ecc85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.203220 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829bd903-9e3e-4378-9c56-6a6375d26422-kube-api-access-tz4t2" (OuterVolumeSpecName: "kube-api-access-tz4t2") pod "829bd903-9e3e-4378-9c56-6a6375d26422" (UID: "829bd903-9e3e-4378-9c56-6a6375d26422"). InnerVolumeSpecName "kube-api-access-tz4t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.204089 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d299e74-bffe-4773-a33a-d615606ecc85-kube-api-access-7748t" (OuterVolumeSpecName: "kube-api-access-7748t") pod "9d299e74-bffe-4773-a33a-d615606ecc85" (UID: "9d299e74-bffe-4773-a33a-d615606ecc85"). InnerVolumeSpecName "kube-api-access-7748t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.204807 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b17ba7e1-deb0-4a31-8874-ee1e6686a3f8-kube-api-access-wdjq5" (OuterVolumeSpecName: "kube-api-access-wdjq5") pod "b17ba7e1-deb0-4a31-8874-ee1e6686a3f8" (UID: "b17ba7e1-deb0-4a31-8874-ee1e6686a3f8"). InnerVolumeSpecName "kube-api-access-wdjq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.211612 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10-kube-api-access-nkrz5" (OuterVolumeSpecName: "kube-api-access-nkrz5") pod "1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10" (UID: "1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10"). InnerVolumeSpecName "kube-api-access-nkrz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.218139 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e4de98-7d7d-43af-a76b-abf1774c620d-kube-api-access-6fmm8" (OuterVolumeSpecName: "kube-api-access-6fmm8") pod "c9e4de98-7d7d-43af-a76b-abf1774c620d" (UID: "c9e4de98-7d7d-43af-a76b-abf1774c620d"). InnerVolumeSpecName "kube-api-access-6fmm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.235736 4871 scope.go:117] "RemoveContainer" containerID="ae69967c6d8b30455e625a813a9b799217766f4ad4774d0faf2565e9178b0ee4" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.238727 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="54142d12-b692-4acd-83f7-e3c029b69d3a" containerName="galera" containerID="cri-o://38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e" gracePeriod=30 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.266086 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-internal-tls-certs\") pod \"878958ac-64b8-434b-88ee-f23d3c486765\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.266203 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/878958ac-64b8-434b-88ee-f23d3c486765-logs\") pod \"878958ac-64b8-434b-88ee-f23d3c486765\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.266358 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79bzm\" (UniqueName: \"kubernetes.io/projected/878958ac-64b8-434b-88ee-f23d3c486765-kube-api-access-79bzm\") pod \"878958ac-64b8-434b-88ee-f23d3c486765\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.266383 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-scripts\") pod \"878958ac-64b8-434b-88ee-f23d3c486765\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.266730 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-combined-ca-bundle\") pod \"878958ac-64b8-434b-88ee-f23d3c486765\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.266813 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-public-tls-certs\") pod \"878958ac-64b8-434b-88ee-f23d3c486765\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.266908 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-config-data\") pod \"878958ac-64b8-434b-88ee-f23d3c486765\" (UID: \"878958ac-64b8-434b-88ee-f23d3c486765\") " Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.267826 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkrz5\" (UniqueName: \"kubernetes.io/projected/1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10-kube-api-access-nkrz5\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.267855 4871 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d299e74-bffe-4773-a33a-d615606ecc85-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.267866 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz4t2\" (UniqueName: \"kubernetes.io/projected/829bd903-9e3e-4378-9c56-6a6375d26422-kube-api-access-tz4t2\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.267877 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdjq5\" (UniqueName: \"kubernetes.io/projected/b17ba7e1-deb0-4a31-8874-ee1e6686a3f8-kube-api-access-wdjq5\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.267886 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fmm8\" (UniqueName: \"kubernetes.io/projected/c9e4de98-7d7d-43af-a76b-abf1774c620d-kube-api-access-6fmm8\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.267895 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d299e74-bffe-4773-a33a-d615606ecc85-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.267904 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.267915 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7748t\" (UniqueName: \"kubernetes.io/projected/9d299e74-bffe-4773-a33a-d615606ecc85-kube-api-access-7748t\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.267924 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.283256 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/878958ac-64b8-434b-88ee-f23d3c486765-logs" (OuterVolumeSpecName: "logs") pod "878958ac-64b8-434b-88ee-f23d3c486765" (UID: "878958ac-64b8-434b-88ee-f23d3c486765"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.302514 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.307660 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.315241 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-scripts" (OuterVolumeSpecName: "scripts") pod "878958ac-64b8-434b-88ee-f23d3c486765" (UID: "878958ac-64b8-434b-88ee-f23d3c486765"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.318264 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/878958ac-64b8-434b-88ee-f23d3c486765-kube-api-access-79bzm" (OuterVolumeSpecName: "kube-api-access-79bzm") pod "878958ac-64b8-434b-88ee-f23d3c486765" (UID: "878958ac-64b8-434b-88ee-f23d3c486765"). InnerVolumeSpecName "kube-api-access-79bzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.320719 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.320867 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="ovn-northd" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.349631 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-677599d776-jjqd2" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:58048->10.217.0.158:9311: read: connection reset by peer" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.350107 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-677599d776-jjqd2" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:58064->10.217.0.158:9311: read: connection reset by peer" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.379034 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79bzm\" (UniqueName: \"kubernetes.io/projected/878958ac-64b8-434b-88ee-f23d3c486765-kube-api-access-79bzm\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.379061 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.379070 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/878958ac-64b8-434b-88ee-f23d3c486765-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.441539 4871 generic.go:334] "Generic (PLEG): container finished" podID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerID="2bf5bc8c358110db521839c0ee5aeb6e45d6fb3a5fe434cba3a0bfc2e3f361c7" exitCode=0 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.441672 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68dde0ec-a90a-4b25-b329-ec645d32671f","Type":"ContainerDied","Data":"2bf5bc8c358110db521839c0ee5aeb6e45d6fb3a5fe434cba3a0bfc2e3f361c7"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.441706 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"68dde0ec-a90a-4b25-b329-ec645d32671f","Type":"ContainerDied","Data":"61ebb153ce635bf338df4f6177e97ab9c1321079b1e7c75b3d6c18c4f67d2324"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.441718 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61ebb153ce635bf338df4f6177e97ab9c1321079b1e7c75b3d6c18c4f67d2324" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.450197 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "829bd903-9e3e-4378-9c56-6a6375d26422" (UID: "829bd903-9e3e-4378-9c56-6a6375d26422"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.459028 4871 generic.go:334] "Generic (PLEG): container finished" podID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerID="a6bcbbe481eb19cf659465c0d4fcb0fa01acc7646b80f2eaea5e2ac2187ebf78" exitCode=0 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.459182 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f1ac984d-67d5-4bd1-b3c2-ef99976a8708","Type":"ContainerDied","Data":"a6bcbbe481eb19cf659465c0d4fcb0fa01acc7646b80f2eaea5e2ac2187ebf78"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.459218 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f1ac984d-67d5-4bd1-b3c2-ef99976a8708","Type":"ContainerDied","Data":"b1edbe024bdd7b1061a46862e4599c8dafd94febc9e77e27c28eb7de224466f9"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.459231 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1edbe024bdd7b1061a46862e4599c8dafd94febc9e77e27c28eb7de224466f9" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.466064 4871 generic.go:334] "Generic (PLEG): container finished" podID="9d299e74-bffe-4773-a33a-d615606ecc85" containerID="b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee" exitCode=0 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.466156 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9d299e74-bffe-4773-a33a-d615606ecc85","Type":"ContainerDied","Data":"b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.466195 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9d299e74-bffe-4773-a33a-d615606ecc85","Type":"ContainerDied","Data":"a4d1e01c9e86249541f40c22a00c33c59fa5c2d70e5a004c774be47711b9bc1a"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.466268 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.479929 4871 generic.go:334] "Generic (PLEG): container finished" podID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerID="83a8879c5c8652cd944b767603633f0434212ef09dde9ebec772c3c1fe6093a1" exitCode=0 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.481373 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23b48c5c-8fd6-4578-bc71-3f7b47860649","Type":"ContainerDied","Data":"83a8879c5c8652cd944b767603633f0434212ef09dde9ebec772c3c1fe6093a1"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.482548 4871 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.499456 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "878958ac-64b8-434b-88ee-f23d3c486765" (UID: "878958ac-64b8-434b-88ee-f23d3c486765"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.502418 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderf74a-account-delete-mvrf4" event={"ID":"1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10","Type":"ContainerDied","Data":"e33c1e5fc4fb38799f00d2cd45f268d07a9d843029bb931d15df8f5ed4a7cf65"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.502520 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderf74a-account-delete-mvrf4" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.509070 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"829bd903-9e3e-4378-9c56-6a6375d26422","Type":"ContainerDied","Data":"e2258cefc7f4d38b866c89108dd159557ee09c8dadfdfebb4aa912845fdf0cdc"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.509171 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.525421 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapid46f-account-delete-8z75z" event={"ID":"ca98e71f-5147-4bb7-bd76-e4bb11cfcf42","Type":"ContainerStarted","Data":"655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.525668 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapid46f-account-delete-8z75z" podUID="ca98e71f-5147-4bb7-bd76-e4bb11cfcf42" containerName="mariadb-account-delete" containerID="cri-o://655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda" gracePeriod=30 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.547455 4871 generic.go:334] "Generic (PLEG): container finished" podID="878958ac-64b8-434b-88ee-f23d3c486765" containerID="858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c" exitCode=0 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.547682 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7567f46b5d-kb4h9" event={"ID":"878958ac-64b8-434b-88ee-f23d3c486765","Type":"ContainerDied","Data":"858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.547718 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7567f46b5d-kb4h9" event={"ID":"878958ac-64b8-434b-88ee-f23d3c486765","Type":"ContainerDied","Data":"dd20e5029ebb0b145df9d9a14d23475bba65506ab1c015e2e51ef4f3933650bd"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.547843 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.553996 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapid46f-account-delete-8z75z" podStartSLOduration=6.553980288 podStartE2EDuration="6.553980288s" podCreationTimestamp="2025-10-07 22:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:32:58.550671711 +0000 UTC m=+1452.353369784" watchObservedRunningTime="2025-10-07 22:32:58.553980288 +0000 UTC m=+1452.356678361" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.556777 4871 generic.go:334] "Generic (PLEG): container finished" podID="f602e068-0d3e-4f32-91ca-3ec36c965a7c" containerID="cd6faea6c3b696458cd8465298585cf7714badd7a13ee982186f92932560d733" exitCode=0 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.556873 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f602e068-0d3e-4f32-91ca-3ec36c965a7c","Type":"ContainerDied","Data":"cd6faea6c3b696458cd8465298585cf7714badd7a13ee982186f92932560d733"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.563084 4871 generic.go:334] "Generic (PLEG): container finished" podID="e71edefc-c645-4e74-8077-42df9e751ce0" containerID="457a86497864d11e171b8cdea112cccb3d4066b22a76ceab718229ee7ea9a56c" exitCode=0 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.563126 4871 generic.go:334] "Generic (PLEG): container finished" podID="e71edefc-c645-4e74-8077-42df9e751ce0" containerID="b2140ff400665da608cd7a343da108840518d021eb0a870d34ed4a85d295f742" exitCode=0 Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.563212 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancea79e-account-delete-jv6lz" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.576268 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementf2b9-account-delete-llr4x" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.586272 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.586645 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.586752 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerDied","Data":"457a86497864d11e171b8cdea112cccb3d4066b22a76ceab718229ee7ea9a56c"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.586835 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerDied","Data":"b2140ff400665da608cd7a343da108840518d021eb0a870d34ed4a85d295f742"} Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.589358 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone6eb9-account-delete-7xq7t" Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.610103 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.610284 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.613433 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.614252 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.614280 4871 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.614558 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.619788 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.619839 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.622396 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9d299e74-bffe-4773-a33a-d615606ecc85" (UID: "9d299e74-bffe-4773-a33a-d615606ecc85"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.631966 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-config-data" (OuterVolumeSpecName: "config-data") pod "c9e4de98-7d7d-43af-a76b-abf1774c620d" (UID: "c9e4de98-7d7d-43af-a76b-abf1774c620d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.646325 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "829bd903-9e3e-4378-9c56-6a6375d26422" (UID: "829bd903-9e3e-4378-9c56-6a6375d26422"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.669773 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "829bd903-9e3e-4378-9c56-6a6375d26422" (UID: "829bd903-9e3e-4378-9c56-6a6375d26422"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.692561 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stmjr\" (UniqueName: \"kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr\") pod \"keystone6eb9-account-delete-7xq7t\" (UID: \"27550e09-71f4-4ed9-96c1-fd40f3f10cfb\") " pod="openstack/keystone6eb9-account-delete-7xq7t" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.694154 4871 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.694505 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.694518 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.694528 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bd903-9e3e-4378-9c56-6a6375d26422-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.705829 4871 projected.go:194] Error preparing data for projected volume kube-api-access-stmjr for pod openstack/keystone6eb9-account-delete-7xq7t: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 07 22:32:58 crc kubenswrapper[4871]: E1007 22:32:58.705920 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr podName:27550e09-71f4-4ed9-96c1-fd40f3f10cfb nodeName:}" failed. No retries permitted until 2025-10-07 22:33:00.705890432 +0000 UTC m=+1454.508588505 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-stmjr" (UniqueName: "kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr") pod "keystone6eb9-account-delete-7xq7t" (UID: "27550e09-71f4-4ed9-96c1-fd40f3f10cfb") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.719118 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "c9e4de98-7d7d-43af-a76b-abf1774c620d" (UID: "c9e4de98-7d7d-43af-a76b-abf1774c620d"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.732600 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d299e74-bffe-4773-a33a-d615606ecc85" (UID: "9d299e74-bffe-4773-a33a-d615606ecc85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.777077 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "878958ac-64b8-434b-88ee-f23d3c486765" (UID: "878958ac-64b8-434b-88ee-f23d3c486765"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.778992 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "c9e4de98-7d7d-43af-a76b-abf1774c620d" (UID: "c9e4de98-7d7d-43af-a76b-abf1774c620d"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.786939 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8vjd9" podUID="22210cb6-0be2-4f3d-9e04-d36274391d54" containerName="ovn-controller" probeResult="failure" output="command timed out" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.802917 4871 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.802953 4871 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.802964 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.802973 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.813976 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9e4de98-7d7d-43af-a76b-abf1774c620d" (UID: "c9e4de98-7d7d-43af-a76b-abf1774c620d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.814168 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9d299e74-bffe-4773-a33a-d615606ecc85" (UID: "9d299e74-bffe-4773-a33a-d615606ecc85"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.830404 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-config-data" (OuterVolumeSpecName: "config-data") pod "878958ac-64b8-434b-88ee-f23d3c486765" (UID: "878958ac-64b8-434b-88ee-f23d3c486765"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.847310 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data" (OuterVolumeSpecName: "config-data") pod "9d299e74-bffe-4773-a33a-d615606ecc85" (UID: "9d299e74-bffe-4773-a33a-d615606ecc85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.876025 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8vjd9" podUID="22210cb6-0be2-4f3d-9e04-d36274391d54" containerName="ovn-controller" probeResult="failure" output=< Oct 07 22:32:58 crc kubenswrapper[4871]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Oct 07 22:32:58 crc kubenswrapper[4871]: > Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.876085 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "878958ac-64b8-434b-88ee-f23d3c486765" (UID: "878958ac-64b8-434b-88ee-f23d3c486765"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.903910 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.903952 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e4de98-7d7d-43af-a76b-abf1774c620d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.903966 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.903979 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878958ac-64b8-434b-88ee-f23d3c486765-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.903991 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d299e74-bffe-4773-a33a-d615606ecc85-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:58 crc kubenswrapper[4871]: I1007 22:32:58.972311 4871 scope.go:117] "RemoveContainer" containerID="b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.016626 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90be0dd0-a9e9-44c9-8142-4ad0d82646ce" path="/var/lib/kubelet/pods/90be0dd0-a9e9-44c9-8142-4ad0d82646ce/volumes" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.017400 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6594e86-288e-4da2-8057-71b4fa8ed8b3" path="/var/lib/kubelet/pods/b6594e86-288e-4da2-8057-71b4fa8ed8b3/volumes" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.018295 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1207599-6a0d-4019-83c0-8c0d6bed143b" path="/var/lib/kubelet/pods/e1207599-6a0d-4019-83c0-8c0d6bed143b/volumes" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.032169 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.068396 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.084967 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.088876 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.088936 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="fd68bd87-c8ef-4214-a9b2-299efdcb956e" containerName="nova-cell1-conductor-conductor" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.143309 4871 scope.go:117] "RemoveContainer" containerID="8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.215697 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-public-tls-certs\") pod \"68dde0ec-a90a-4b25-b329-ec645d32671f\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.215905 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68dde0ec-a90a-4b25-b329-ec645d32671f-logs\") pod \"68dde0ec-a90a-4b25-b329-ec645d32671f\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.215961 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-internal-tls-certs\") pod \"68dde0ec-a90a-4b25-b329-ec645d32671f\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.215992 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxq8l\" (UniqueName: \"kubernetes.io/projected/68dde0ec-a90a-4b25-b329-ec645d32671f-kube-api-access-xxq8l\") pod \"68dde0ec-a90a-4b25-b329-ec645d32671f\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.216022 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-combined-ca-bundle\") pod \"68dde0ec-a90a-4b25-b329-ec645d32671f\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.216115 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-config-data\") pod \"68dde0ec-a90a-4b25-b329-ec645d32671f\" (UID: \"68dde0ec-a90a-4b25-b329-ec645d32671f\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.221435 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68dde0ec-a90a-4b25-b329-ec645d32671f-kube-api-access-xxq8l" (OuterVolumeSpecName: "kube-api-access-xxq8l") pod "68dde0ec-a90a-4b25-b329-ec645d32671f" (UID: "68dde0ec-a90a-4b25-b329-ec645d32671f"). InnerVolumeSpecName "kube-api-access-xxq8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.223347 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68dde0ec-a90a-4b25-b329-ec645d32671f-logs" (OuterVolumeSpecName: "logs") pod "68dde0ec-a90a-4b25-b329-ec645d32671f" (UID: "68dde0ec-a90a-4b25-b329-ec645d32671f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.248067 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-config-data" (OuterVolumeSpecName: "config-data") pod "68dde0ec-a90a-4b25-b329-ec645d32671f" (UID: "68dde0ec-a90a-4b25-b329-ec645d32671f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.252206 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68dde0ec-a90a-4b25-b329-ec645d32671f" (UID: "68dde0ec-a90a-4b25-b329-ec645d32671f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.301959 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "68dde0ec-a90a-4b25-b329-ec645d32671f" (UID: "68dde0ec-a90a-4b25-b329-ec645d32671f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.306310 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "68dde0ec-a90a-4b25-b329-ec645d32671f" (UID: "68dde0ec-a90a-4b25-b329-ec645d32671f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.320376 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68dde0ec-a90a-4b25-b329-ec645d32671f-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.320401 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.320412 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxq8l\" (UniqueName: \"kubernetes.io/projected/68dde0ec-a90a-4b25-b329-ec645d32671f-kube-api-access-xxq8l\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.320421 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.320432 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.320440 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68dde0ec-a90a-4b25-b329-ec645d32671f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.408404 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="655d7d7e1b7fb65057965cba90062d42b1f499d86be0b09af648ef03eef592b5" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.417506 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="655d7d7e1b7fb65057965cba90062d42b1f499d86be0b09af648ef03eef592b5" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.421182 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="655d7d7e1b7fb65057965cba90062d42b1f499d86be0b09af648ef03eef592b5" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.421281 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="1e61b0b6-83cc-48b0-9f3d-b9b607faa830" containerName="nova-cell0-conductor-conductor" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.629843 4871 generic.go:334] "Generic (PLEG): container finished" podID="e71edefc-c645-4e74-8077-42df9e751ce0" containerID="f02b2caa39b480dc045d7e7262686751f44a5f1546d461810390f4aa68b67b32" exitCode=0 Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.629925 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerDied","Data":"f02b2caa39b480dc045d7e7262686751f44a5f1546d461810390f4aa68b67b32"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.633619 4871 generic.go:334] "Generic (PLEG): container finished" podID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerID="91eb51510e831d42f41a70585ae844bb55e88150c93eae1571a50da925c20bf3" exitCode=0 Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.633669 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d02631b4-f3a6-48d5-a43e-c90a48622ec2","Type":"ContainerDied","Data":"91eb51510e831d42f41a70585ae844bb55e88150c93eae1571a50da925c20bf3"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.633691 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d02631b4-f3a6-48d5-a43e-c90a48622ec2","Type":"ContainerDied","Data":"5f7c928bf87134511100f2455772d75ebb4dade15c1424ad2a868ec3ac15b223"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.633702 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f7c928bf87134511100f2455772d75ebb4dade15c1424ad2a868ec3ac15b223" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.646841 4871 generic.go:334] "Generic (PLEG): container finished" podID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerID="f1bb4a9d4f25de043a799d48cedbb3a8b9bba20dd97ad863253a098ddb8d0065" exitCode=0 Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.646965 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5544d45d85-dqp9x" event={"ID":"f9973055-9eec-4599-9d2b-ccef87f6a481","Type":"ContainerDied","Data":"f1bb4a9d4f25de043a799d48cedbb3a8b9bba20dd97ad863253a098ddb8d0065"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.695344 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f602e068-0d3e-4f32-91ca-3ec36c965a7c","Type":"ContainerDied","Data":"47f00885da8c1cfefb3e371701cc2c79f37aaf083da95cc13e2c2698888acca7"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.695390 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47f00885da8c1cfefb3e371701cc2c79f37aaf083da95cc13e2c2698888acca7" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.757415 4871 generic.go:334] "Generic (PLEG): container finished" podID="1e61b0b6-83cc-48b0-9f3d-b9b607faa830" containerID="655d7d7e1b7fb65057965cba90062d42b1f499d86be0b09af648ef03eef592b5" exitCode=0 Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.758083 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1e61b0b6-83cc-48b0-9f3d-b9b607faa830","Type":"ContainerDied","Data":"655d7d7e1b7fb65057965cba90062d42b1f499d86be0b09af648ef03eef592b5"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.780265 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23b48c5c-8fd6-4578-bc71-3f7b47860649","Type":"ContainerDied","Data":"03f7e7bbcdc1cbea796f537923a2f519f4899845d8d312992fcd10ea14e3d1dd"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.780349 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03f7e7bbcdc1cbea796f537923a2f519f4899845d8d312992fcd10ea14e3d1dd" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.787840 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderf74a-account-delete-mvrf4"] Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.790925 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.798868 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.820276 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinderf74a-account-delete-mvrf4"] Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.822270 4871 generic.go:334] "Generic (PLEG): container finished" podID="8724dd55-b509-4056-b324-531ec57c3a7e" containerID="2a05385f6943057613877aa85b019b47371856ca80b0e282c3f1b443a5093210" exitCode=0 Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.822340 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" event={"ID":"8724dd55-b509-4056-b324-531ec57c3a7e","Type":"ContainerDied","Data":"2a05385f6943057613877aa85b019b47371856ca80b0e282c3f1b443a5093210"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.828101 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone6eb9-account-delete-7xq7t" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.841900 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementf2b9-account-delete-llr4x"] Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.856297 4871 scope.go:117] "RemoveContainer" containerID="b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.856739 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.857979 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee\": container with ID starting with b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee not found: ID does not exist" containerID="b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.858040 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee"} err="failed to get container status \"b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee\": rpc error: code = NotFound desc = could not find container \"b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee\": container with ID starting with b4d8379aa459c3a3934b94c5e1b4f40afcb9639e0861a0c0ef8753508a7863ee not found: ID does not exist" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.858076 4871 scope.go:117] "RemoveContainer" containerID="8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.858296 4871 generic.go:334] "Generic (PLEG): container finished" podID="c750808e-de6d-442b-82ed-d3498b036d91" containerID="e6f0d276a1846b36e45c6f4882a0d1c7b6023b4bc31e6f161d4fcc4d02c212f3" exitCode=0 Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.858401 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.858406 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677599d776-jjqd2" event={"ID":"c750808e-de6d-442b-82ed-d3498b036d91","Type":"ContainerDied","Data":"e6f0d276a1846b36e45c6f4882a0d1c7b6023b4bc31e6f161d4fcc4d02c212f3"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.858468 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677599d776-jjqd2" event={"ID":"c750808e-de6d-442b-82ed-d3498b036d91","Type":"ContainerDied","Data":"94b95aca5eb11093877243f356cbb8de96ca88027ee3aa0bdfdcca7dab2af178"} Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.858488 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94b95aca5eb11093877243f356cbb8de96ca88027ee3aa0bdfdcca7dab2af178" Oct 07 22:32:59 crc kubenswrapper[4871]: E1007 22:32:59.863519 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047\": container with ID starting with 8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047 not found: ID does not exist" containerID="8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.863669 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047"} err="failed to get container status \"8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047\": rpc error: code = NotFound desc = could not find container \"8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047\": container with ID starting with 8c1b00451a2992b7b9a694d4c649f4798d2d86f13b28da6cade062b841a8e047 not found: ID does not exist" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.863707 4871 scope.go:117] "RemoveContainer" containerID="f988c8bf25a955bb983d8bdd6ca560c46444fc402fe7eefe2c5fdbd7ae80b4c5" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.866206 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.875281 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementf2b9-account-delete-llr4x"] Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.937806 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancea79e-account-delete-jv6lz"] Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963228 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-memcached-tls-certs\") pod \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963316 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-logs\") pod \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963340 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-scripts\") pod \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963371 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kolla-config\") pod \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963397 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgf2v\" (UniqueName: \"kubernetes.io/projected/23b48c5c-8fd6-4578-bc71-3f7b47860649-kube-api-access-xgf2v\") pod \"23b48c5c-8fd6-4578-bc71-3f7b47860649\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963419 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963447 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4rcl\" (UniqueName: \"kubernetes.io/projected/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-kube-api-access-r4rcl\") pod \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963556 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-httpd-run\") pod \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963578 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-nova-metadata-tls-certs\") pod \"23b48c5c-8fd6-4578-bc71-3f7b47860649\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963601 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-config-data\") pod \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963620 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-public-tls-certs\") pod \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963661 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-config-data\") pod \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963687 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-logs\") pod \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.963939 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b48c5c-8fd6-4578-bc71-3f7b47860649-logs\") pod \"23b48c5c-8fd6-4578-bc71-3f7b47860649\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964134 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-internal-tls-certs\") pod \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964161 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-config-data\") pod \"23b48c5c-8fd6-4578-bc71-3f7b47860649\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964185 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-config-data\") pod \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964207 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-combined-ca-bundle\") pod \"23b48c5c-8fd6-4578-bc71-3f7b47860649\" (UID: \"23b48c5c-8fd6-4578-bc71-3f7b47860649\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964225 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-scripts\") pod \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964249 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-combined-ca-bundle\") pod \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964268 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d22dl\" (UniqueName: \"kubernetes.io/projected/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kube-api-access-d22dl\") pod \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964371 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-combined-ca-bundle\") pod \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\" (UID: \"f602e068-0d3e-4f32-91ca-3ec36c965a7c\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964399 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9djzg\" (UniqueName: \"kubernetes.io/projected/d02631b4-f3a6-48d5-a43e-c90a48622ec2-kube-api-access-9djzg\") pod \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964426 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-combined-ca-bundle\") pod \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\" (UID: \"d02631b4-f3a6-48d5-a43e-c90a48622ec2\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964442 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.964496 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-httpd-run\") pod \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\" (UID: \"f1ac984d-67d5-4bd1-b3c2-ef99976a8708\") " Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.965903 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-logs" (OuterVolumeSpecName: "logs") pod "d02631b4-f3a6-48d5-a43e-c90a48622ec2" (UID: "d02631b4-f3a6-48d5-a43e-c90a48622ec2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.967963 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f1ac984d-67d5-4bd1-b3c2-ef99976a8708" (UID: "f1ac984d-67d5-4bd1-b3c2-ef99976a8708"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.969128 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23b48c5c-8fd6-4578-bc71-3f7b47860649-logs" (OuterVolumeSpecName: "logs") pod "23b48c5c-8fd6-4578-bc71-3f7b47860649" (UID: "23b48c5c-8fd6-4578-bc71-3f7b47860649"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.973836 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d02631b4-f3a6-48d5-a43e-c90a48622ec2" (UID: "d02631b4-f3a6-48d5-a43e-c90a48622ec2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.974615 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "f602e068-0d3e-4f32-91ca-3ec36c965a7c" (UID: "f602e068-0d3e-4f32-91ca-3ec36c965a7c"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.976163 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-logs" (OuterVolumeSpecName: "logs") pod "f1ac984d-67d5-4bd1-b3c2-ef99976a8708" (UID: "f1ac984d-67d5-4bd1-b3c2-ef99976a8708"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.982005 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-config-data" (OuterVolumeSpecName: "config-data") pod "f602e068-0d3e-4f32-91ca-3ec36c965a7c" (UID: "f602e068-0d3e-4f32-91ca-3ec36c965a7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.984425 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kube-api-access-d22dl" (OuterVolumeSpecName: "kube-api-access-d22dl") pod "f602e068-0d3e-4f32-91ca-3ec36c965a7c" (UID: "f602e068-0d3e-4f32-91ca-3ec36c965a7c"). InnerVolumeSpecName "kube-api-access-d22dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.989063 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d02631b4-f3a6-48d5-a43e-c90a48622ec2-kube-api-access-9djzg" (OuterVolumeSpecName: "kube-api-access-9djzg") pod "d02631b4-f3a6-48d5-a43e-c90a48622ec2" (UID: "d02631b4-f3a6-48d5-a43e-c90a48622ec2"). InnerVolumeSpecName "kube-api-access-9djzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.993420 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:32:59 crc kubenswrapper[4871]: I1007 22:32:59.994901 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glancea79e-account-delete-jv6lz"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.001950 4871 scope.go:117] "RemoveContainer" containerID="1b18488a0daf1c9ccb071f5a1f222406f2336c9c403fbfe6cbfc92c70b53edf6" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.006618 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-scripts" (OuterVolumeSpecName: "scripts") pod "f1ac984d-67d5-4bd1-b3c2-ef99976a8708" (UID: "f1ac984d-67d5-4bd1-b3c2-ef99976a8708"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.007485 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.012336 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "f1ac984d-67d5-4bd1-b3c2-ef99976a8708" (UID: "f1ac984d-67d5-4bd1-b3c2-ef99976a8708"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.018610 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.023734 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "d02631b4-f3a6-48d5-a43e-c90a48622ec2" (UID: "d02631b4-f3a6-48d5-a43e-c90a48622ec2"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.026614 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-scripts" (OuterVolumeSpecName: "scripts") pod "d02631b4-f3a6-48d5-a43e-c90a48622ec2" (UID: "d02631b4-f3a6-48d5-a43e-c90a48622ec2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.027911 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-kube-api-access-r4rcl" (OuterVolumeSpecName: "kube-api-access-r4rcl") pod "f1ac984d-67d5-4bd1-b3c2-ef99976a8708" (UID: "f1ac984d-67d5-4bd1-b3c2-ef99976a8708"). InnerVolumeSpecName "kube-api-access-r4rcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.035962 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.044151 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23b48c5c-8fd6-4578-bc71-3f7b47860649-kube-api-access-xgf2v" (OuterVolumeSpecName: "kube-api-access-xgf2v") pod "23b48c5c-8fd6-4578-bc71-3f7b47860649" (UID: "23b48c5c-8fd6-4578-bc71-3f7b47860649"). InnerVolumeSpecName "kube-api-access-xgf2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.051330 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.054591 4871 scope.go:117] "RemoveContainer" containerID="858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.067353 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.075815 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.077418 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.077466 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d22dl\" (UniqueName: \"kubernetes.io/projected/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kube-api-access-d22dl\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.077482 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9djzg\" (UniqueName: \"kubernetes.io/projected/d02631b4-f3a6-48d5-a43e-c90a48622ec2-kube-api-access-9djzg\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.077533 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.077550 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.077567 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.077654 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.078179 4871 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.078271 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgf2v\" (UniqueName: \"kubernetes.io/projected/23b48c5c-8fd6-4578-bc71-3f7b47860649-kube-api-access-xgf2v\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.078358 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.078424 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4rcl\" (UniqueName: \"kubernetes.io/projected/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-kube-api-access-r4rcl\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.078486 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.078539 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f602e068-0d3e-4f32-91ca-3ec36c965a7c-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.078591 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d02631b4-f3a6-48d5-a43e-c90a48622ec2-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.078641 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23b48c5c-8fd6-4578-bc71-3f7b47860649-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.110514 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.112999 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.121409 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.137183 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.143039 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23b48c5c-8fd6-4578-bc71-3f7b47860649" (UID: "23b48c5c-8fd6-4578-bc71-3f7b47860649"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.147384 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-config-data" (OuterVolumeSpecName: "config-data") pod "23b48c5c-8fd6-4578-bc71-3f7b47860649" (UID: "23b48c5c-8fd6-4578-bc71-3f7b47860649"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.148918 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1ac984d-67d5-4bd1-b3c2-ef99976a8708" (UID: "f1ac984d-67d5-4bd1-b3c2-ef99976a8708"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.157422 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f602e068-0d3e-4f32-91ca-3ec36c965a7c" (UID: "f602e068-0d3e-4f32-91ca-3ec36c965a7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.158482 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d02631b4-f3a6-48d5-a43e-c90a48622ec2" (UID: "d02631b4-f3a6-48d5-a43e-c90a48622ec2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.164435 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.171122 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.179331 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.179905 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-run-httpd\") pod \"e71edefc-c645-4e74-8077-42df9e751ce0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.179946 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ptvb\" (UniqueName: \"kubernetes.io/projected/e71edefc-c645-4e74-8077-42df9e751ce0-kube-api-access-6ptvb\") pod \"e71edefc-c645-4e74-8077-42df9e751ce0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.179988 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-internal-tls-certs\") pod \"c750808e-de6d-442b-82ed-d3498b036d91\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180010 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data-custom\") pod \"c750808e-de6d-442b-82ed-d3498b036d91\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180033 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-scripts\") pod \"e71edefc-c645-4e74-8077-42df9e751ce0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180049 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-log-httpd\") pod \"e71edefc-c645-4e74-8077-42df9e751ce0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180077 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-ceilometer-tls-certs\") pod \"e71edefc-c645-4e74-8077-42df9e751ce0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180199 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96hnw\" (UniqueName: \"kubernetes.io/projected/c750808e-de6d-442b-82ed-d3498b036d91-kube-api-access-96hnw\") pod \"c750808e-de6d-442b-82ed-d3498b036d91\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180235 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data\") pod \"c750808e-de6d-442b-82ed-d3498b036d91\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180261 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-config-data\") pod \"e71edefc-c645-4e74-8077-42df9e751ce0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180283 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-combined-ca-bundle\") pod \"e71edefc-c645-4e74-8077-42df9e751ce0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180302 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data-custom\") pod \"8724dd55-b509-4056-b324-531ec57c3a7e\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180338 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8724dd55-b509-4056-b324-531ec57c3a7e-logs\") pod \"8724dd55-b509-4056-b324-531ec57c3a7e\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180372 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-public-tls-certs\") pod \"c750808e-de6d-442b-82ed-d3498b036d91\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180395 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data\") pod \"8724dd55-b509-4056-b324-531ec57c3a7e\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180449 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c750808e-de6d-442b-82ed-d3498b036d91-logs\") pod \"c750808e-de6d-442b-82ed-d3498b036d91\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180474 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-sg-core-conf-yaml\") pod \"e71edefc-c645-4e74-8077-42df9e751ce0\" (UID: \"e71edefc-c645-4e74-8077-42df9e751ce0\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180498 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9sj8\" (UniqueName: \"kubernetes.io/projected/8724dd55-b509-4056-b324-531ec57c3a7e-kube-api-access-h9sj8\") pod \"8724dd55-b509-4056-b324-531ec57c3a7e\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180522 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-combined-ca-bundle\") pod \"8724dd55-b509-4056-b324-531ec57c3a7e\" (UID: \"8724dd55-b509-4056-b324-531ec57c3a7e\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.180594 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-combined-ca-bundle\") pod \"c750808e-de6d-442b-82ed-d3498b036d91\" (UID: \"c750808e-de6d-442b-82ed-d3498b036d91\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.181540 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.181599 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.181614 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.181627 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.181639 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.181695 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.181707 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.193522 4871 scope.go:117] "RemoveContainer" containerID="ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc" Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.194320 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.194403 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data podName:dd77ac8e-36b6-4e16-869a-3ce6bff223ac nodeName:}" failed. No retries permitted until 2025-10-07 22:33:08.19438137 +0000 UTC m=+1461.997079433 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data") pod "rabbitmq-server-0" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac") : configmap "rabbitmq-config-data" not found Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.195224 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c750808e-de6d-442b-82ed-d3498b036d91-logs" (OuterVolumeSpecName: "logs") pod "c750808e-de6d-442b-82ed-d3498b036d91" (UID: "c750808e-de6d-442b-82ed-d3498b036d91"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.197938 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e71edefc-c645-4e74-8077-42df9e751ce0" (UID: "e71edefc-c645-4e74-8077-42df9e751ce0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.198767 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e71edefc-c645-4e74-8077-42df9e751ce0" (UID: "e71edefc-c645-4e74-8077-42df9e751ce0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.200127 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8724dd55-b509-4056-b324-531ec57c3a7e-logs" (OuterVolumeSpecName: "logs") pod "8724dd55-b509-4056-b324-531ec57c3a7e" (UID: "8724dd55-b509-4056-b324-531ec57c3a7e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.203443 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-scripts" (OuterVolumeSpecName: "scripts") pod "e71edefc-c645-4e74-8077-42df9e751ce0" (UID: "e71edefc-c645-4e74-8077-42df9e751ce0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.209554 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-config-data" (OuterVolumeSpecName: "config-data") pod "d02631b4-f3a6-48d5-a43e-c90a48622ec2" (UID: "d02631b4-f3a6-48d5-a43e-c90a48622ec2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.224498 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "23b48c5c-8fd6-4578-bc71-3f7b47860649" (UID: "23b48c5c-8fd6-4578-bc71-3f7b47860649"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.242111 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8724dd55-b509-4056-b324-531ec57c3a7e" (UID: "8724dd55-b509-4056-b324-531ec57c3a7e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.242118 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c750808e-de6d-442b-82ed-d3498b036d91" (UID: "c750808e-de6d-442b-82ed-d3498b036d91"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.242128 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c750808e-de6d-442b-82ed-d3498b036d91-kube-api-access-96hnw" (OuterVolumeSpecName: "kube-api-access-96hnw") pod "c750808e-de6d-442b-82ed-d3498b036d91" (UID: "c750808e-de6d-442b-82ed-d3498b036d91"). InnerVolumeSpecName "kube-api-access-96hnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.242220 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8724dd55-b509-4056-b324-531ec57c3a7e-kube-api-access-h9sj8" (OuterVolumeSpecName: "kube-api-access-h9sj8") pod "8724dd55-b509-4056-b324-531ec57c3a7e" (UID: "8724dd55-b509-4056-b324-531ec57c3a7e"). InnerVolumeSpecName "kube-api-access-h9sj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.242303 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e71edefc-c645-4e74-8077-42df9e751ce0-kube-api-access-6ptvb" (OuterVolumeSpecName: "kube-api-access-6ptvb") pod "e71edefc-c645-4e74-8077-42df9e751ce0" (UID: "e71edefc-c645-4e74-8077-42df9e751ce0"). InnerVolumeSpecName "kube-api-access-6ptvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.260473 4871 scope.go:117] "RemoveContainer" containerID="858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c" Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.263091 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c\": container with ID starting with 858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c not found: ID does not exist" containerID="858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.263163 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c"} err="failed to get container status \"858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c\": rpc error: code = NotFound desc = could not find container \"858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c\": container with ID starting with 858f3564f7ad49e1cfdc6b90f18c3c40cd03bb6c095821761db6e3bd8e70047c not found: ID does not exist" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.263194 4871 scope.go:117] "RemoveContainer" containerID="ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc" Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.263833 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc\": container with ID starting with ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc not found: ID does not exist" containerID="ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.263893 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc"} err="failed to get container status \"ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc\": rpc error: code = NotFound desc = could not find container \"ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc\": container with ID starting with ff28bb789cdf4490d8fd5c6b2c4f57c9c40aaa4bbce9d05ca29bb6af437f1cfc not found: ID does not exist" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.282458 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data-custom\") pod \"f9973055-9eec-4599-9d2b-ccef87f6a481\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.282570 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2vdj\" (UniqueName: \"kubernetes.io/projected/f9973055-9eec-4599-9d2b-ccef87f6a481-kube-api-access-s2vdj\") pod \"f9973055-9eec-4599-9d2b-ccef87f6a481\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.282703 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data\") pod \"f9973055-9eec-4599-9d2b-ccef87f6a481\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.282723 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-combined-ca-bundle\") pod \"f9973055-9eec-4599-9d2b-ccef87f6a481\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.282881 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9973055-9eec-4599-9d2b-ccef87f6a481-logs\") pod \"f9973055-9eec-4599-9d2b-ccef87f6a481\" (UID: \"f9973055-9eec-4599-9d2b-ccef87f6a481\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283313 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c750808e-de6d-442b-82ed-d3498b036d91-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283326 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9sj8\" (UniqueName: \"kubernetes.io/projected/8724dd55-b509-4056-b324-531ec57c3a7e-kube-api-access-h9sj8\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283337 4871 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23b48c5c-8fd6-4578-bc71-3f7b47860649-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283346 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283354 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283365 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ptvb\" (UniqueName: \"kubernetes.io/projected/e71edefc-c645-4e74-8077-42df9e751ce0-kube-api-access-6ptvb\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283373 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283382 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283390 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e71edefc-c645-4e74-8077-42df9e751ce0-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283400 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96hnw\" (UniqueName: \"kubernetes.io/projected/c750808e-de6d-442b-82ed-d3498b036d91-kube-api-access-96hnw\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283411 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283420 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8724dd55-b509-4056-b324-531ec57c3a7e-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.283818 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9973055-9eec-4599-9d2b-ccef87f6a481-logs" (OuterVolumeSpecName: "logs") pod "f9973055-9eec-4599-9d2b-ccef87f6a481" (UID: "f9973055-9eec-4599-9d2b-ccef87f6a481"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.288304 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.302211 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9973055-9eec-4599-9d2b-ccef87f6a481-kube-api-access-s2vdj" (OuterVolumeSpecName: "kube-api-access-s2vdj") pod "f9973055-9eec-4599-9d2b-ccef87f6a481" (UID: "f9973055-9eec-4599-9d2b-ccef87f6a481"). InnerVolumeSpecName "kube-api-access-s2vdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.316234 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f9973055-9eec-4599-9d2b-ccef87f6a481" (UID: "f9973055-9eec-4599-9d2b-ccef87f6a481"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.365665 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-config-data" (OuterVolumeSpecName: "config-data") pod "f1ac984d-67d5-4bd1-b3c2-ef99976a8708" (UID: "f1ac984d-67d5-4bd1-b3c2-ef99976a8708"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.375011 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d02631b4-f3a6-48d5-a43e-c90a48622ec2" (UID: "d02631b4-f3a6-48d5-a43e-c90a48622ec2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.385661 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-combined-ca-bundle\") pod \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.385851 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-config-data\") pod \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.385889 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78gk8\" (UniqueName: \"kubernetes.io/projected/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-kube-api-access-78gk8\") pod \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\" (UID: \"1e61b0b6-83cc-48b0-9f3d-b9b607faa830\") " Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.386357 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d02631b4-f3a6-48d5-a43e-c90a48622ec2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.386842 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.386857 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9973055-9eec-4599-9d2b-ccef87f6a481-logs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.386873 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.386886 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2vdj\" (UniqueName: \"kubernetes.io/projected/f9973055-9eec-4599-9d2b-ccef87f6a481-kube-api-access-s2vdj\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.421169 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data" (OuterVolumeSpecName: "config-data") pod "c750808e-de6d-442b-82ed-d3498b036d91" (UID: "c750808e-de6d-442b-82ed-d3498b036d91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.421199 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-kube-api-access-78gk8" (OuterVolumeSpecName: "kube-api-access-78gk8") pod "1e61b0b6-83cc-48b0-9f3d-b9b607faa830" (UID: "1e61b0b6-83cc-48b0-9f3d-b9b607faa830"). InnerVolumeSpecName "kube-api-access-78gk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.490362 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78gk8\" (UniqueName: \"kubernetes.io/projected/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-kube-api-access-78gk8\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.490403 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.525958 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e61b0b6-83cc-48b0-9f3d-b9b607faa830" (UID: "1e61b0b6-83cc-48b0-9f3d-b9b607faa830"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.539641 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "f602e068-0d3e-4f32-91ca-3ec36c965a7c" (UID: "f602e068-0d3e-4f32-91ca-3ec36c965a7c"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.598980 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8724dd55-b509-4056-b324-531ec57c3a7e" (UID: "8724dd55-b509-4056-b324-531ec57c3a7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.599370 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.600048 4871 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f602e068-0d3e-4f32-91ca-3ec36c965a7c-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.600164 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.603448 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c750808e-de6d-442b-82ed-d3498b036d91" (UID: "c750808e-de6d-442b-82ed-d3498b036d91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.689392 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e71edefc-c645-4e74-8077-42df9e751ce0" (UID: "e71edefc-c645-4e74-8077-42df9e751ce0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.703049 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.703081 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.732012 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9973055-9eec-4599-9d2b-ccef87f6a481" (UID: "f9973055-9eec-4599-9d2b-ccef87f6a481"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.737270 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c750808e-de6d-442b-82ed-d3498b036d91" (UID: "c750808e-de6d-442b-82ed-d3498b036d91"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.800133 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f1ac984d-67d5-4bd1-b3c2-ef99976a8708" (UID: "f1ac984d-67d5-4bd1-b3c2-ef99976a8708"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.805461 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stmjr\" (UniqueName: \"kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr\") pod \"keystone6eb9-account-delete-7xq7t\" (UID: \"27550e09-71f4-4ed9-96c1-fd40f3f10cfb\") " pod="openstack/keystone6eb9-account-delete-7xq7t" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.805677 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.805735 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1ac984d-67d5-4bd1-b3c2-ef99976a8708-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.805748 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.809576 4871 projected.go:194] Error preparing data for projected volume kube-api-access-stmjr for pod openstack/keystone6eb9-account-delete-7xq7t: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.809651 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr podName:27550e09-71f4-4ed9-96c1-fd40f3f10cfb nodeName:}" failed. No retries permitted until 2025-10-07 22:33:04.809618005 +0000 UTC m=+1458.612316078 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-stmjr" (UniqueName: "kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr") pod "keystone6eb9-account-delete-7xq7t" (UID: "27550e09-71f4-4ed9-96c1-fd40f3f10cfb") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.820251 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data" (OuterVolumeSpecName: "config-data") pod "8724dd55-b509-4056-b324-531ec57c3a7e" (UID: "8724dd55-b509-4056-b324-531ec57c3a7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.829990 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-config-data" (OuterVolumeSpecName: "config-data") pod "1e61b0b6-83cc-48b0-9f3d-b9b607faa830" (UID: "1e61b0b6-83cc-48b0-9f3d-b9b607faa830"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.834936 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c750808e-de6d-442b-82ed-d3498b036d91" (UID: "c750808e-de6d-442b-82ed-d3498b036d91"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.844913 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data" (OuterVolumeSpecName: "config-data") pod "f9973055-9eec-4599-9d2b-ccef87f6a481" (UID: "f9973055-9eec-4599-9d2b-ccef87f6a481"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.857283 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.861150 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "e71edefc-c645-4e74-8077-42df9e751ce0" (UID: "e71edefc-c645-4e74-8077-42df9e751ce0"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.864645 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e71edefc-c645-4e74-8077-42df9e751ce0" (UID: "e71edefc-c645-4e74-8077-42df9e751ce0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.874763 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.879246 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-config-data" (OuterVolumeSpecName: "config-data") pod "e71edefc-c645-4e74-8077-42df9e751ce0" (UID: "e71edefc-c645-4e74-8077-42df9e751ce0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.880913 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5544d45d85-dqp9x" event={"ID":"f9973055-9eec-4599-9d2b-ccef87f6a481","Type":"ContainerDied","Data":"f9ba6f90866aa6a695fac71c4be8863f248545b55cda441d59d486660156a5b0"} Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.880936 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5544d45d85-dqp9x" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.880975 4871 scope.go:117] "RemoveContainer" containerID="f1bb4a9d4f25de043a799d48cedbb3a8b9bba20dd97ad863253a098ddb8d0065" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.890564 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.890726 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e71edefc-c645-4e74-8077-42df9e751ce0","Type":"ContainerDied","Data":"33c6f7c23038063d0da214f2c17dddbf223e200146c8e1e89dee21bab58b13a5"} Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.890744 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.895390 4871 generic.go:334] "Generic (PLEG): container finished" podID="6af0bcb3-3803-4174-af06-c6841610c62a" containerID="1987ababc57bf8086a0c131045faa2ea781c10edf009a458861520b942f65314" exitCode=0 Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.895513 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cd55cdc75-kj6jh" event={"ID":"6af0bcb3-3803-4174-af06-c6841610c62a","Type":"ContainerDied","Data":"1987ababc57bf8086a0c131045faa2ea781c10edf009a458861520b942f65314"} Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.895665 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.896709 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 22:33:00 crc kubenswrapper[4871]: E1007 22:33:00.896813 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ab7d1def-9358-44f0-9245-e333fb60bac4" containerName="nova-scheduler-scheduler" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.908056 4871 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.908140 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.908150 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71edefc-c645-4e74-8077-42df9e751ce0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.908161 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c750808e-de6d-442b-82ed-d3498b036d91-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.908183 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8724dd55-b509-4056-b324-531ec57c3a7e-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.908193 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e61b0b6-83cc-48b0-9f3d-b9b607faa830-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.908202 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9973055-9eec-4599-9d2b-ccef87f6a481-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.918015 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" event={"ID":"8724dd55-b509-4056-b324-531ec57c3a7e","Type":"ContainerDied","Data":"8844a5d2ddf30eba5f81c5e76504bc6ea6fb435a7f5ff614b2673ba68c45ae7c"} Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.918446 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bc6797fd8-gflqx" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.920329 4871 scope.go:117] "RemoveContainer" containerID="26e20164e3525af55f84ab8ba8cc88a36fbcc0795cd362085e6b25b516d314d7" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.928441 4871 generic.go:334] "Generic (PLEG): container finished" podID="fd68bd87-c8ef-4214-a9b2-299efdcb956e" containerID="f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63" exitCode=0 Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.928856 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fd68bd87-c8ef-4214-a9b2-299efdcb956e","Type":"ContainerDied","Data":"f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63"} Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.929033 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fd68bd87-c8ef-4214-a9b2-299efdcb956e","Type":"ContainerDied","Data":"cef18de0b674288b076c26646fef70938e4878c638247a3cbf2e8da910d5acd1"} Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.929221 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.941886 4871 generic.go:334] "Generic (PLEG): container finished" podID="54142d12-b692-4acd-83f7-e3c029b69d3a" containerID="38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e" exitCode=0 Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.942043 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54142d12-b692-4acd-83f7-e3c029b69d3a","Type":"ContainerDied","Data":"38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e"} Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.942085 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54142d12-b692-4acd-83f7-e3c029b69d3a","Type":"ContainerDied","Data":"cbbf66cd67f9bbcaa232bbb87909cf67f8090f8e8a6840833b437243a2d72314"} Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.942223 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.953319 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1e61b0b6-83cc-48b0-9f3d-b9b607faa830","Type":"ContainerDied","Data":"35140a7bd0b66f48c5c839c75acf092a674587aee56df1429be08e5ab25f5fce"} Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.953493 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.979879 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-677599d776-jjqd2" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.980768 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.980991 4871 scope.go:117] "RemoveContainer" containerID="457a86497864d11e171b8cdea112cccb3d4066b22a76ceab718229ee7ea9a56c" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.982421 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.982476 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.982506 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone6eb9-account-delete-7xq7t" Oct 07 22:33:00 crc kubenswrapper[4871]: I1007 22:33:00.983351 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.010895 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10" path="/var/lib/kubelet/pods/1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10/volumes" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.011501 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="361d6d85-f8c7-4643-95c4-29741f91a179" path="/var/lib/kubelet/pods/361d6d85-f8c7-4643-95c4-29741f91a179/volumes" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.011906 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-default\") pod \"54142d12-b692-4acd-83f7-e3c029b69d3a\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.011991 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-operator-scripts\") pod \"54142d12-b692-4acd-83f7-e3c029b69d3a\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012071 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-combined-ca-bundle\") pod \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012134 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb5vn\" (UniqueName: \"kubernetes.io/projected/54142d12-b692-4acd-83f7-e3c029b69d3a-kube-api-access-zb5vn\") pod \"54142d12-b692-4acd-83f7-e3c029b69d3a\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012173 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"54142d12-b692-4acd-83f7-e3c029b69d3a\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012210 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmq8z\" (UniqueName: \"kubernetes.io/projected/fd68bd87-c8ef-4214-a9b2-299efdcb956e-kube-api-access-pmq8z\") pod \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012265 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-config-data\") pod \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\" (UID: \"fd68bd87-c8ef-4214-a9b2-299efdcb956e\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012294 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-secrets\") pod \"54142d12-b692-4acd-83f7-e3c029b69d3a\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012325 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-generated\") pod \"54142d12-b692-4acd-83f7-e3c029b69d3a\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012386 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-combined-ca-bundle\") pod \"54142d12-b692-4acd-83f7-e3c029b69d3a\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012415 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-galera-tls-certs\") pod \"54142d12-b692-4acd-83f7-e3c029b69d3a\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.012462 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-kolla-config\") pod \"54142d12-b692-4acd-83f7-e3c029b69d3a\" (UID: \"54142d12-b692-4acd-83f7-e3c029b69d3a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.013898 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "54142d12-b692-4acd-83f7-e3c029b69d3a" (UID: "54142d12-b692-4acd-83f7-e3c029b69d3a"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.014229 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "54142d12-b692-4acd-83f7-e3c029b69d3a" (UID: "54142d12-b692-4acd-83f7-e3c029b69d3a"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.014408 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54142d12-b692-4acd-83f7-e3c029b69d3a" (UID: "54142d12-b692-4acd-83f7-e3c029b69d3a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.017568 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" path="/var/lib/kubelet/pods/68dde0ec-a90a-4b25-b329-ec645d32671f/volumes" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.018730 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "54142d12-b692-4acd-83f7-e3c029b69d3a" (UID: "54142d12-b692-4acd-83f7-e3c029b69d3a"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.019369 4871 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.019387 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.019397 4871 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54142d12-b692-4acd-83f7-e3c029b69d3a-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.019412 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/54142d12-b692-4acd-83f7-e3c029b69d3a-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.020503 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="829bd903-9e3e-4378-9c56-6a6375d26422" path="/var/lib/kubelet/pods/829bd903-9e3e-4378-9c56-6a6375d26422/volumes" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.023148 4871 scope.go:117] "RemoveContainer" containerID="b1bc98f764cc4f92e14878e9fcdb1b85ceb45251562f19cc562511bd4c01c7c5" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.024825 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" path="/var/lib/kubelet/pods/9d299e74-bffe-4773-a33a-d615606ecc85/volumes" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.026537 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b17ba7e1-deb0-4a31-8874-ee1e6686a3f8" path="/var/lib/kubelet/pods/b17ba7e1-deb0-4a31-8874-ee1e6686a3f8/volumes" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.027211 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e4de98-7d7d-43af-a76b-abf1774c620d" path="/var/lib/kubelet/pods/c9e4de98-7d7d-43af-a76b-abf1774c620d/volumes" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.032103 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd68bd87-c8ef-4214-a9b2-299efdcb956e-kube-api-access-pmq8z" (OuterVolumeSpecName: "kube-api-access-pmq8z") pod "fd68bd87-c8ef-4214-a9b2-299efdcb956e" (UID: "fd68bd87-c8ef-4214-a9b2-299efdcb956e"). InnerVolumeSpecName "kube-api-access-pmq8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.052388 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-secrets" (OuterVolumeSpecName: "secrets") pod "54142d12-b692-4acd-83f7-e3c029b69d3a" (UID: "54142d12-b692-4acd-83f7-e3c029b69d3a"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.052729 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-config-data" (OuterVolumeSpecName: "config-data") pod "fd68bd87-c8ef-4214-a9b2-299efdcb956e" (UID: "fd68bd87-c8ef-4214-a9b2-299efdcb956e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.064954 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54142d12-b692-4acd-83f7-e3c029b69d3a-kube-api-access-zb5vn" (OuterVolumeSpecName: "kube-api-access-zb5vn") pod "54142d12-b692-4acd-83f7-e3c029b69d3a" (UID: "54142d12-b692-4acd-83f7-e3c029b69d3a"). InnerVolumeSpecName "kube-api-access-zb5vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.073083 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.074304 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5544d45d85-dqp9x"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.074346 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-5544d45d85-dqp9x"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.074364 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.087667 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "mysql-db") pod "54142d12-b692-4acd-83f7-e3c029b69d3a" (UID: "54142d12-b692-4acd-83f7-e3c029b69d3a"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.089620 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd68bd87-c8ef-4214-a9b2-299efdcb956e" (UID: "fd68bd87-c8ef-4214-a9b2-299efdcb956e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.090029 4871 scope.go:117] "RemoveContainer" containerID="f02b2caa39b480dc045d7e7262686751f44a5f1546d461810390f4aa68b67b32" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.098008 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54142d12-b692-4acd-83f7-e3c029b69d3a" (UID: "54142d12-b692-4acd-83f7-e3c029b69d3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.101258 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.117710 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.123298 4871 scope.go:117] "RemoveContainer" containerID="b2140ff400665da608cd7a343da108840518d021eb0a870d34ed4a85d295f742" Oct 07 22:33:01 crc kubenswrapper[4871]: E1007 22:33:01.124655 4871 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.124686 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmq8z\" (UniqueName: \"kubernetes.io/projected/fd68bd87-c8ef-4214-a9b2-299efdcb956e-kube-api-access-pmq8z\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: E1007 22:33:01.124739 4871 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data podName:1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb nodeName:}" failed. No retries permitted until 2025-10-07 22:33:09.12471112 +0000 UTC m=+1462.927409183 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data") pod "rabbitmq-cell1-server-0" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb") : configmap "rabbitmq-cell1-config-data" not found Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.124773 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.124822 4871 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.124833 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.124847 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd68bd87-c8ef-4214-a9b2-299efdcb956e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.124859 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb5vn\" (UniqueName: \"kubernetes.io/projected/54142d12-b692-4acd-83f7-e3c029b69d3a-kube-api-access-zb5vn\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.124885 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.125970 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.132337 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6bc6797fd8-gflqx"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.140983 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-6bc6797fd8-gflqx"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.147848 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.158975 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-677599d776-jjqd2"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.171585 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "54142d12-b692-4acd-83f7-e3c029b69d3a" (UID: "54142d12-b692-4acd-83f7-e3c029b69d3a"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.176158 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-677599d776-jjqd2"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.193598 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.193686 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.193721 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.196249 4871 scope.go:117] "RemoveContainer" containerID="2a05385f6943057613877aa85b019b47371856ca80b0e282c3f1b443a5093210" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.207430 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.214743 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.223203 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.234004 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-public-tls-certs\") pod \"6af0bcb3-3803-4174-af06-c6841610c62a\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.234052 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-fernet-keys\") pod \"6af0bcb3-3803-4174-af06-c6841610c62a\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.234636 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t7gq\" (UniqueName: \"kubernetes.io/projected/6af0bcb3-3803-4174-af06-c6841610c62a-kube-api-access-4t7gq\") pod \"6af0bcb3-3803-4174-af06-c6841610c62a\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.235031 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-credential-keys\") pod \"6af0bcb3-3803-4174-af06-c6841610c62a\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.235058 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-config-data\") pod \"6af0bcb3-3803-4174-af06-c6841610c62a\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.235130 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-internal-tls-certs\") pod \"6af0bcb3-3803-4174-af06-c6841610c62a\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.235218 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-combined-ca-bundle\") pod \"6af0bcb3-3803-4174-af06-c6841610c62a\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.235626 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-scripts\") pod \"6af0bcb3-3803-4174-af06-c6841610c62a\" (UID: \"6af0bcb3-3803-4174-af06-c6841610c62a\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.243189 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6af0bcb3-3803-4174-af06-c6841610c62a" (UID: "6af0bcb3-3803-4174-af06-c6841610c62a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.247299 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-scripts" (OuterVolumeSpecName: "scripts") pod "6af0bcb3-3803-4174-af06-c6841610c62a" (UID: "6af0bcb3-3803-4174-af06-c6841610c62a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.250346 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone6eb9-account-delete-7xq7t"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.257347 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6af0bcb3-3803-4174-af06-c6841610c62a-kube-api-access-4t7gq" (OuterVolumeSpecName: "kube-api-access-4t7gq") pod "6af0bcb3-3803-4174-af06-c6841610c62a" (UID: "6af0bcb3-3803-4174-af06-c6841610c62a"). InnerVolumeSpecName "kube-api-access-4t7gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.257359 4871 scope.go:117] "RemoveContainer" containerID="d4cf3f49a66e3bfb1a2f0116e08e29835a98701bf050feff3371d723047d7b5c" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.257676 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6af0bcb3-3803-4174-af06-c6841610c62a" (UID: "6af0bcb3-3803-4174-af06-c6841610c62a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.259546 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.259579 4871 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/54142d12-b692-4acd-83f7-e3c029b69d3a-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.260996 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone6eb9-account-delete-7xq7t"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.274831 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.282705 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-config-data" (OuterVolumeSpecName: "config-data") pod "6af0bcb3-3803-4174-af06-c6841610c62a" (UID: "6af0bcb3-3803-4174-af06-c6841610c62a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.283596 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.286180 4871 scope.go:117] "RemoveContainer" containerID="f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.298695 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.298748 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.302321 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6af0bcb3-3803-4174-af06-c6841610c62a" (UID: "6af0bcb3-3803-4174-af06-c6841610c62a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.312077 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6af0bcb3-3803-4174-af06-c6841610c62a" (UID: "6af0bcb3-3803-4174-af06-c6841610c62a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.313298 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.320506 4871 scope.go:117] "RemoveContainer" containerID="f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63" Oct 07 22:33:01 crc kubenswrapper[4871]: E1007 22:33:01.320958 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63\": container with ID starting with f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63 not found: ID does not exist" containerID="f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.320993 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63"} err="failed to get container status \"f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63\": rpc error: code = NotFound desc = could not find container \"f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63\": container with ID starting with f53e20e0b91c4e14d444febe549b01b5987cc4cac0e150342299f107e300dc63 not found: ID does not exist" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.321018 4871 scope.go:117] "RemoveContainer" containerID="38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.323141 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.350494 4871 scope.go:117] "RemoveContainer" containerID="ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.363484 4871 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.363518 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.363528 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.363541 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stmjr\" (UniqueName: \"kubernetes.io/projected/27550e09-71f4-4ed9-96c1-fd40f3f10cfb-kube-api-access-stmjr\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.363552 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.363561 4871 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.363571 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.363580 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t7gq\" (UniqueName: \"kubernetes.io/projected/6af0bcb3-3803-4174-af06-c6841610c62a-kube-api-access-4t7gq\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.369317 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6af0bcb3-3803-4174-af06-c6841610c62a" (UID: "6af0bcb3-3803-4174-af06-c6841610c62a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.465459 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af0bcb3-3803-4174-af06-c6841610c62a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.585351 4871 scope.go:117] "RemoveContainer" containerID="38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e" Oct 07 22:33:01 crc kubenswrapper[4871]: E1007 22:33:01.586343 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e\": container with ID starting with 38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e not found: ID does not exist" containerID="38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.586421 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e"} err="failed to get container status \"38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e\": rpc error: code = NotFound desc = could not find container \"38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e\": container with ID starting with 38369fb0a58fa797ec4fdd207312db1ecd884fbe4571211523210045792c759e not found: ID does not exist" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.586467 4871 scope.go:117] "RemoveContainer" containerID="ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232" Oct 07 22:33:01 crc kubenswrapper[4871]: E1007 22:33:01.589293 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232\": container with ID starting with ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232 not found: ID does not exist" containerID="ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.589330 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232"} err="failed to get container status \"ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232\": rpc error: code = NotFound desc = could not find container \"ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232\": container with ID starting with ee0997c30fcf34847ec7044a69cdf541fc8141f10e406979d233376a43118232 not found: ID does not exist" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.589362 4871 scope.go:117] "RemoveContainer" containerID="655d7d7e1b7fb65057965cba90062d42b1f499d86be0b09af648ef03eef592b5" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.846176 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.909969 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7cc4f647bc-9lrsh" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.153:9696/\": dial tcp 10.217.0.153:9696: connect: connection refused" Oct 07 22:33:01 crc kubenswrapper[4871]: E1007 22:33:01.953540 4871 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 07 22:33:01 crc kubenswrapper[4871]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-07T22:32:54Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 07 22:33:01 crc kubenswrapper[4871]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 07 22:33:01 crc kubenswrapper[4871]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-8vjd9" message=< Oct 07 22:33:01 crc kubenswrapper[4871]: Exiting ovn-controller (1) [FAILED] Oct 07 22:33:01 crc kubenswrapper[4871]: Killing ovn-controller (1) [ OK ] Oct 07 22:33:01 crc kubenswrapper[4871]: Killing ovn-controller (1) with SIGKILL [ OK ] Oct 07 22:33:01 crc kubenswrapper[4871]: 2025-10-07T22:32:54Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 07 22:33:01 crc kubenswrapper[4871]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 07 22:33:01 crc kubenswrapper[4871]: > Oct 07 22:33:01 crc kubenswrapper[4871]: E1007 22:33:01.953592 4871 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 07 22:33:01 crc kubenswrapper[4871]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-07T22:32:54Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 07 22:33:01 crc kubenswrapper[4871]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 07 22:33:01 crc kubenswrapper[4871]: > pod="openstack/ovn-controller-8vjd9" podUID="22210cb6-0be2-4f3d-9e04-d36274391d54" containerName="ovn-controller" containerID="cri-o://ae1cc42e1fe96bd37dd99ecb252d123aafccbfb8f26e86ac08399eae8a5f9b5d" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.953652 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-8vjd9" podUID="22210cb6-0be2-4f3d-9e04-d36274391d54" containerName="ovn-controller" containerID="cri-o://ae1cc42e1fe96bd37dd99ecb252d123aafccbfb8f26e86ac08399eae8a5f9b5d" gracePeriod=22 Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.978888 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-erlang-cookie\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.978984 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-pod-info\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.979015 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.979035 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-erlang-cookie-secret\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.979615 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.979859 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-tls\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.979897 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77kcp\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-kube-api-access-77kcp\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.979930 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-plugins-conf\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.979965 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-plugins\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.979995 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.980014 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-confd\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.980055 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-server-conf\") pod \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\" (UID: \"dd77ac8e-36b6-4e16-869a-3ce6bff223ac\") " Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.980545 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.980720 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.980740 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.981105 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.994126 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.995732 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-pod-info" (OuterVolumeSpecName: "pod-info") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.996038 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.996159 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-kube-api-access-77kcp" (OuterVolumeSpecName: "kube-api-access-77kcp") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "kube-api-access-77kcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:01 crc kubenswrapper[4871]: I1007 22:33:01.996554 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.004350 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fb7d0be2-8b44-44ed-948d-11e1932d27b4/ovn-northd/0.log" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.004425 4871 generic.go:334] "Generic (PLEG): container finished" podID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerID="691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303" exitCode=139 Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.004488 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fb7d0be2-8b44-44ed-948d-11e1932d27b4","Type":"ContainerDied","Data":"691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303"} Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.008238 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cd55cdc75-kj6jh" event={"ID":"6af0bcb3-3803-4174-af06-c6841610c62a","Type":"ContainerDied","Data":"72047d6988570f2778705a14820615c6601e749d79b7cf666481ed8423c52cb0"} Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.008283 4871 scope.go:117] "RemoveContainer" containerID="1987ababc57bf8086a0c131045faa2ea781c10edf009a458861520b942f65314" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.008289 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cd55cdc75-kj6jh" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.016235 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8vjd9_22210cb6-0be2-4f3d-9e04-d36274391d54/ovn-controller/0.log" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.016528 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8vjd9" event={"ID":"22210cb6-0be2-4f3d-9e04-d36274391d54","Type":"ContainerDied","Data":"ae1cc42e1fe96bd37dd99ecb252d123aafccbfb8f26e86ac08399eae8a5f9b5d"} Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.016407 4871 generic.go:334] "Generic (PLEG): container finished" podID="22210cb6-0be2-4f3d-9e04-d36274391d54" containerID="ae1cc42e1fe96bd37dd99ecb252d123aafccbfb8f26e86ac08399eae8a5f9b5d" exitCode=137 Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.017660 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data" (OuterVolumeSpecName: "config-data") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.021175 4871 generic.go:334] "Generic (PLEG): container finished" podID="dd77ac8e-36b6-4e16-869a-3ce6bff223ac" containerID="fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb" exitCode=0 Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.021245 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.021237 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dd77ac8e-36b6-4e16-869a-3ce6bff223ac","Type":"ContainerDied","Data":"fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb"} Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.021304 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dd77ac8e-36b6-4e16-869a-3ce6bff223ac","Type":"ContainerDied","Data":"7ab7035a520d5d9b271e84ffbaa98a26ad959ea341166ea18f363c62cd14e0f4"} Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.040130 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-server-conf" (OuterVolumeSpecName: "server-conf") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.048608 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-cd55cdc75-kj6jh"] Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.058766 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-cd55cdc75-kj6jh"] Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.062954 4871 scope.go:117] "RemoveContainer" containerID="fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.082436 4871 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-pod-info\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.082468 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.082476 4871 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.082487 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.082496 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77kcp\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-kube-api-access-77kcp\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.082504 4871 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.082527 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.082539 4871 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-server-conf\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.089480 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fb7d0be2-8b44-44ed-948d-11e1932d27b4/ovn-northd/0.log" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.089583 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.099902 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.104730 4871 scope.go:117] "RemoveContainer" containerID="ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.123428 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "dd77ac8e-36b6-4e16-869a-3ce6bff223ac" (UID: "dd77ac8e-36b6-4e16-869a-3ce6bff223ac"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.142781 4871 scope.go:117] "RemoveContainer" containerID="fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb" Oct 07 22:33:02 crc kubenswrapper[4871]: E1007 22:33:02.145942 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb\": container with ID starting with fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb not found: ID does not exist" containerID="fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.145996 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb"} err="failed to get container status \"fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb\": rpc error: code = NotFound desc = could not find container \"fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb\": container with ID starting with fbc619c1c946b3b84f208d771d426d0ab0d086e7d5a3d4cb937f78fe5370d6eb not found: ID does not exist" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.146034 4871 scope.go:117] "RemoveContainer" containerID="ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2" Oct 07 22:33:02 crc kubenswrapper[4871]: E1007 22:33:02.146598 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2\": container with ID starting with ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2 not found: ID does not exist" containerID="ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.146658 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2"} err="failed to get container status \"ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2\": rpc error: code = NotFound desc = could not find container \"ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2\": container with ID starting with ed5072647b4bf1cdad80589731793461a4f9e322daa0703566329eed846066f2 not found: ID does not exist" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.183950 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-scripts\") pod \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.184143 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-metrics-certs-tls-certs\") pod \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.184251 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv797\" (UniqueName: \"kubernetes.io/projected/fb7d0be2-8b44-44ed-948d-11e1932d27b4-kube-api-access-wv797\") pod \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.184351 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-rundir\") pod \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.184395 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-combined-ca-bundle\") pod \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.184424 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-northd-tls-certs\") pod \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.184522 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-config\") pod \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\" (UID: \"fb7d0be2-8b44-44ed-948d-11e1932d27b4\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.185743 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "fb7d0be2-8b44-44ed-948d-11e1932d27b4" (UID: "fb7d0be2-8b44-44ed-948d-11e1932d27b4"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.185874 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-scripts" (OuterVolumeSpecName: "scripts") pod "fb7d0be2-8b44-44ed-948d-11e1932d27b4" (UID: "fb7d0be2-8b44-44ed-948d-11e1932d27b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.186351 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-config" (OuterVolumeSpecName: "config") pod "fb7d0be2-8b44-44ed-948d-11e1932d27b4" (UID: "fb7d0be2-8b44-44ed-948d-11e1932d27b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.186661 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.186677 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.186687 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb7d0be2-8b44-44ed-948d-11e1932d27b4-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.186695 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.186705 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dd77ac8e-36b6-4e16-869a-3ce6bff223ac-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.189282 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb7d0be2-8b44-44ed-948d-11e1932d27b4-kube-api-access-wv797" (OuterVolumeSpecName: "kube-api-access-wv797") pod "fb7d0be2-8b44-44ed-948d-11e1932d27b4" (UID: "fb7d0be2-8b44-44ed-948d-11e1932d27b4"). InnerVolumeSpecName "kube-api-access-wv797". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.264239 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "fb7d0be2-8b44-44ed-948d-11e1932d27b4" (UID: "fb7d0be2-8b44-44ed-948d-11e1932d27b4"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.267916 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb7d0be2-8b44-44ed-948d-11e1932d27b4" (UID: "fb7d0be2-8b44-44ed-948d-11e1932d27b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.287751 4871 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.288210 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv797\" (UniqueName: \"kubernetes.io/projected/fb7d0be2-8b44-44ed-948d-11e1932d27b4-kube-api-access-wv797\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.288225 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.292673 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "fb7d0be2-8b44-44ed-948d-11e1932d27b4" (UID: "fb7d0be2-8b44-44ed-948d-11e1932d27b4"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.311548 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8vjd9_22210cb6-0be2-4f3d-9e04-d36274391d54/ovn-controller/0.log" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.311637 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8vjd9" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389427 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run-ovn\") pod \"22210cb6-0be2-4f3d-9e04-d36274391d54\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389513 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-log-ovn\") pod \"22210cb6-0be2-4f3d-9e04-d36274391d54\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389550 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run\") pod \"22210cb6-0be2-4f3d-9e04-d36274391d54\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389606 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f22hp\" (UniqueName: \"kubernetes.io/projected/22210cb6-0be2-4f3d-9e04-d36274391d54-kube-api-access-f22hp\") pod \"22210cb6-0be2-4f3d-9e04-d36274391d54\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389658 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-combined-ca-bundle\") pod \"22210cb6-0be2-4f3d-9e04-d36274391d54\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389695 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22210cb6-0be2-4f3d-9e04-d36274391d54-scripts\") pod \"22210cb6-0be2-4f3d-9e04-d36274391d54\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389721 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-ovn-controller-tls-certs\") pod \"22210cb6-0be2-4f3d-9e04-d36274391d54\" (UID: \"22210cb6-0be2-4f3d-9e04-d36274391d54\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389772 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run" (OuterVolumeSpecName: "var-run") pod "22210cb6-0be2-4f3d-9e04-d36274391d54" (UID: "22210cb6-0be2-4f3d-9e04-d36274391d54"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389862 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "22210cb6-0be2-4f3d-9e04-d36274391d54" (UID: "22210cb6-0be2-4f3d-9e04-d36274391d54"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389884 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "22210cb6-0be2-4f3d-9e04-d36274391d54" (UID: "22210cb6-0be2-4f3d-9e04-d36274391d54"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.390162 4871 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.390183 4871 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.390197 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7d0be2-8b44-44ed-948d-11e1932d27b4-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.390209 4871 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/22210cb6-0be2-4f3d-9e04-d36274391d54-var-run\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.397066 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22210cb6-0be2-4f3d-9e04-d36274391d54-scripts" (OuterVolumeSpecName: "scripts") pod "22210cb6-0be2-4f3d-9e04-d36274391d54" (UID: "22210cb6-0be2-4f3d-9e04-d36274391d54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.389681 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.429610 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22210cb6-0be2-4f3d-9e04-d36274391d54-kube-api-access-f22hp" (OuterVolumeSpecName: "kube-api-access-f22hp") pod "22210cb6-0be2-4f3d-9e04-d36274391d54" (UID: "22210cb6-0be2-4f3d-9e04-d36274391d54"). InnerVolumeSpecName "kube-api-access-f22hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.436677 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.440006 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22210cb6-0be2-4f3d-9e04-d36274391d54" (UID: "22210cb6-0be2-4f3d-9e04-d36274391d54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.487764 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "22210cb6-0be2-4f3d-9e04-d36274391d54" (UID: "22210cb6-0be2-4f3d-9e04-d36274391d54"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.492298 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f22hp\" (UniqueName: \"kubernetes.io/projected/22210cb6-0be2-4f3d-9e04-d36274391d54-kube-api-access-f22hp\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.492443 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.492553 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22210cb6-0be2-4f3d-9e04-d36274391d54-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.492634 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/22210cb6-0be2-4f3d-9e04-d36274391d54-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.662926 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695582 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g27pd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-kube-api-access-g27pd\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695655 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-server-conf\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695697 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695716 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-erlang-cookie\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695741 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695802 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-tls\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695835 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-plugins-conf\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695900 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-erlang-cookie-secret\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695921 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-confd\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695939 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-pod-info\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.695963 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-plugins\") pod \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\" (UID: \"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb\") " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.696921 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.701644 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.731383 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.731957 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-kube-api-access-g27pd" (OuterVolumeSpecName: "kube-api-access-g27pd") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "kube-api-access-g27pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.732618 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.732879 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.733383 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data" (OuterVolumeSpecName: "config-data") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.733824 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.742104 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-pod-info" (OuterVolumeSpecName: "pod-info") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.764689 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-server-conf" (OuterVolumeSpecName: "server-conf") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.794412 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="c9e4de98-7d7d-43af-a76b-abf1774c620d" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.197:6080/vnc_lite.html\": dial tcp 10.217.0.197:6080: i/o timeout" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798635 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798673 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798686 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798695 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798703 4871 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798712 4871 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798720 4871 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-pod-info\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798728 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798749 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g27pd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-kube-api-access-g27pd\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.798757 4871 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-server-conf\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.819618 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.820142 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" (UID: "1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.900726 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:02 crc kubenswrapper[4871]: I1007 22:33:02.900769 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.000265 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e61b0b6-83cc-48b0-9f3d-b9b607faa830" path="/var/lib/kubelet/pods/1e61b0b6-83cc-48b0-9f3d-b9b607faa830/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.001687 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" path="/var/lib/kubelet/pods/23b48c5c-8fd6-4578-bc71-3f7b47860649/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.002563 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27550e09-71f4-4ed9-96c1-fd40f3f10cfb" path="/var/lib/kubelet/pods/27550e09-71f4-4ed9-96c1-fd40f3f10cfb/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.003407 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54142d12-b692-4acd-83f7-e3c029b69d3a" path="/var/lib/kubelet/pods/54142d12-b692-4acd-83f7-e3c029b69d3a/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.005464 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6af0bcb3-3803-4174-af06-c6841610c62a" path="/var/lib/kubelet/pods/6af0bcb3-3803-4174-af06-c6841610c62a/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.006647 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8724dd55-b509-4056-b324-531ec57c3a7e" path="/var/lib/kubelet/pods/8724dd55-b509-4056-b324-531ec57c3a7e/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.009032 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c750808e-de6d-442b-82ed-d3498b036d91" path="/var/lib/kubelet/pods/c750808e-de6d-442b-82ed-d3498b036d91/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.010487 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" path="/var/lib/kubelet/pods/d02631b4-f3a6-48d5-a43e-c90a48622ec2/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.013084 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd77ac8e-36b6-4e16-869a-3ce6bff223ac" path="/var/lib/kubelet/pods/dd77ac8e-36b6-4e16-869a-3ce6bff223ac/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.015692 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" path="/var/lib/kubelet/pods/e71edefc-c645-4e74-8077-42df9e751ce0/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.018069 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" path="/var/lib/kubelet/pods/f1ac984d-67d5-4bd1-b3c2-ef99976a8708/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.020809 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f602e068-0d3e-4f32-91ca-3ec36c965a7c" path="/var/lib/kubelet/pods/f602e068-0d3e-4f32-91ca-3ec36c965a7c/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.021869 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9973055-9eec-4599-9d2b-ccef87f6a481" path="/var/lib/kubelet/pods/f9973055-9eec-4599-9d2b-ccef87f6a481/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.022950 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd68bd87-c8ef-4214-a9b2-299efdcb956e" path="/var/lib/kubelet/pods/fd68bd87-c8ef-4214-a9b2-299efdcb956e/volumes" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.042356 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fb7d0be2-8b44-44ed-948d-11e1932d27b4/ovn-northd/0.log" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.042462 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fb7d0be2-8b44-44ed-948d-11e1932d27b4","Type":"ContainerDied","Data":"acbdfc581af6868d99263bc705496804406a76fe949cd0e7fcea2854e1352143"} Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.042520 4871 scope.go:117] "RemoveContainer" containerID="a0bcaf69c69657ecf8f36f03f56417c9e0e10c4b3d9151b2ed50c569b93152aa" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.042566 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.048126 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.048164 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb","Type":"ContainerDied","Data":"798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d"} Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.048078 4871 generic.go:334] "Generic (PLEG): container finished" podID="1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" containerID="798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d" exitCode=0 Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.048369 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb","Type":"ContainerDied","Data":"8e9d23a5a31ea2bce14ff1e207052c1d0d2645872d3ba0dbe8e8de2d8a8c684e"} Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.057545 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8vjd9_22210cb6-0be2-4f3d-9e04-d36274391d54/ovn-controller/0.log" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.057761 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8vjd9" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.058044 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8vjd9" event={"ID":"22210cb6-0be2-4f3d-9e04-d36274391d54","Type":"ContainerDied","Data":"a46269d9afc65812aeb877c434e200be29562fb25baecd4383a37363622aa094"} Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.141105 4871 scope.go:117] "RemoveContainer" containerID="691d9792ff0a0b615727cd6fefa43160776ed91fcab31840175f60076c400303" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.176916 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.192772 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.201476 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8vjd9"] Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.223010 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8vjd9"] Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.229899 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.232306 4871 scope.go:117] "RemoveContainer" containerID="798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.240220 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.260144 4871 scope.go:117] "RemoveContainer" containerID="3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.289391 4871 scope.go:117] "RemoveContainer" containerID="798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d" Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.290413 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d\": container with ID starting with 798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d not found: ID does not exist" containerID="798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.290477 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d"} err="failed to get container status \"798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d\": rpc error: code = NotFound desc = could not find container \"798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d\": container with ID starting with 798089ee9af3e224e379795765bc4a2afe9c99292b085e08cd2cf285d9d59c3d not found: ID does not exist" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.290513 4871 scope.go:117] "RemoveContainer" containerID="3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907" Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.290938 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907\": container with ID starting with 3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907 not found: ID does not exist" containerID="3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.290963 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907"} err="failed to get container status \"3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907\": rpc error: code = NotFound desc = could not find container \"3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907\": container with ID starting with 3ec560d2ebb575d6e24b3e15d9fbbb23e0b9375410a46b29b7c9e7838db94907 not found: ID does not exist" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.290977 4871 scope.go:117] "RemoveContainer" containerID="ae1cc42e1fe96bd37dd99ecb252d123aafccbfb8f26e86ac08399eae8a5f9b5d" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.521190 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.581563 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.582029 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.582398 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.582435 4871 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.589235 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.594101 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.595981 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:03 crc kubenswrapper[4871]: E1007 22:33:03.596027 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.630786 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-config-data\") pod \"ab7d1def-9358-44f0-9245-e333fb60bac4\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.630924 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcmfw\" (UniqueName: \"kubernetes.io/projected/ab7d1def-9358-44f0-9245-e333fb60bac4-kube-api-access-rcmfw\") pod \"ab7d1def-9358-44f0-9245-e333fb60bac4\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.630994 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-combined-ca-bundle\") pod \"ab7d1def-9358-44f0-9245-e333fb60bac4\" (UID: \"ab7d1def-9358-44f0-9245-e333fb60bac4\") " Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.641084 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab7d1def-9358-44f0-9245-e333fb60bac4-kube-api-access-rcmfw" (OuterVolumeSpecName: "kube-api-access-rcmfw") pod "ab7d1def-9358-44f0-9245-e333fb60bac4" (UID: "ab7d1def-9358-44f0-9245-e333fb60bac4"). InnerVolumeSpecName "kube-api-access-rcmfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.652956 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-config-data" (OuterVolumeSpecName: "config-data") pod "ab7d1def-9358-44f0-9245-e333fb60bac4" (UID: "ab7d1def-9358-44f0-9245-e333fb60bac4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.660347 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab7d1def-9358-44f0-9245-e333fb60bac4" (UID: "ab7d1def-9358-44f0-9245-e333fb60bac4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.734406 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.734451 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcmfw\" (UniqueName: \"kubernetes.io/projected/ab7d1def-9358-44f0-9245-e333fb60bac4-kube-api-access-rcmfw\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:03 crc kubenswrapper[4871]: I1007 22:33:03.734466 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d1def-9358-44f0-9245-e333fb60bac4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:04 crc kubenswrapper[4871]: I1007 22:33:04.127709 4871 generic.go:334] "Generic (PLEG): container finished" podID="ab7d1def-9358-44f0-9245-e333fb60bac4" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" exitCode=0 Oct 07 22:33:04 crc kubenswrapper[4871]: I1007 22:33:04.127899 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 22:33:04 crc kubenswrapper[4871]: I1007 22:33:04.127933 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ab7d1def-9358-44f0-9245-e333fb60bac4","Type":"ContainerDied","Data":"d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a"} Oct 07 22:33:04 crc kubenswrapper[4871]: I1007 22:33:04.132399 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ab7d1def-9358-44f0-9245-e333fb60bac4","Type":"ContainerDied","Data":"216394cc8d96b3a3aa867cf7ad54dd4656d2395398530a696359975606a58fc1"} Oct 07 22:33:04 crc kubenswrapper[4871]: I1007 22:33:04.132451 4871 scope.go:117] "RemoveContainer" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" Oct 07 22:33:04 crc kubenswrapper[4871]: I1007 22:33:04.163784 4871 scope.go:117] "RemoveContainer" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" Oct 07 22:33:04 crc kubenswrapper[4871]: E1007 22:33:04.164391 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a\": container with ID starting with d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a not found: ID does not exist" containerID="d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a" Oct 07 22:33:04 crc kubenswrapper[4871]: I1007 22:33:04.164431 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a"} err="failed to get container status \"d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a\": rpc error: code = NotFound desc = could not find container \"d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a\": container with ID starting with d6e2afd034642423dc3336914f56d8cc788387fd6a9f21ee1850f26273f8302a not found: ID does not exist" Oct 07 22:33:04 crc kubenswrapper[4871]: I1007 22:33:04.176334 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:33:04 crc kubenswrapper[4871]: I1007 22:33:04.181877 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 22:33:05 crc kubenswrapper[4871]: I1007 22:33:05.001676 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" path="/var/lib/kubelet/pods/1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb/volumes" Oct 07 22:33:05 crc kubenswrapper[4871]: I1007 22:33:05.002775 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22210cb6-0be2-4f3d-9e04-d36274391d54" path="/var/lib/kubelet/pods/22210cb6-0be2-4f3d-9e04-d36274391d54/volumes" Oct 07 22:33:05 crc kubenswrapper[4871]: I1007 22:33:05.003334 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab7d1def-9358-44f0-9245-e333fb60bac4" path="/var/lib/kubelet/pods/ab7d1def-9358-44f0-9245-e333fb60bac4/volumes" Oct 07 22:33:05 crc kubenswrapper[4871]: I1007 22:33:05.004547 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" path="/var/lib/kubelet/pods/fb7d0be2-8b44-44ed-948d-11e1932d27b4/volumes" Oct 07 22:33:05 crc kubenswrapper[4871]: I1007 22:33:05.512185 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:33:05 crc kubenswrapper[4871]: I1007 22:33:05.512268 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:33:08 crc kubenswrapper[4871]: E1007 22:33:08.582668 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:08 crc kubenswrapper[4871]: E1007 22:33:08.583855 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:08 crc kubenswrapper[4871]: E1007 22:33:08.584813 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:08 crc kubenswrapper[4871]: E1007 22:33:08.584923 4871 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" Oct 07 22:33:08 crc kubenswrapper[4871]: E1007 22:33:08.584777 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:08 crc kubenswrapper[4871]: E1007 22:33:08.586731 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:08 crc kubenswrapper[4871]: E1007 22:33:08.589452 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:08 crc kubenswrapper[4871]: E1007 22:33:08.589522 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.133069 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.200882 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-combined-ca-bundle\") pod \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.200929 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-httpd-config\") pod \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.200982 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-public-tls-certs\") pod \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.201032 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-internal-tls-certs\") pod \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.201085 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-ovndb-tls-certs\") pod \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.201119 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pmf8\" (UniqueName: \"kubernetes.io/projected/7116e0fa-30d9-4a8d-9a11-0d57908f268f-kube-api-access-5pmf8\") pod \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.201135 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-config\") pod \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\" (UID: \"7116e0fa-30d9-4a8d-9a11-0d57908f268f\") " Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.210374 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "7116e0fa-30d9-4a8d-9a11-0d57908f268f" (UID: "7116e0fa-30d9-4a8d-9a11-0d57908f268f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.210644 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7116e0fa-30d9-4a8d-9a11-0d57908f268f-kube-api-access-5pmf8" (OuterVolumeSpecName: "kube-api-access-5pmf8") pod "7116e0fa-30d9-4a8d-9a11-0d57908f268f" (UID: "7116e0fa-30d9-4a8d-9a11-0d57908f268f"). InnerVolumeSpecName "kube-api-access-5pmf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.257313 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7116e0fa-30d9-4a8d-9a11-0d57908f268f" (UID: "7116e0fa-30d9-4a8d-9a11-0d57908f268f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.261640 4871 generic.go:334] "Generic (PLEG): container finished" podID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerID="09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0" exitCode=0 Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.261711 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cc4f647bc-9lrsh" event={"ID":"7116e0fa-30d9-4a8d-9a11-0d57908f268f","Type":"ContainerDied","Data":"09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0"} Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.261770 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cc4f647bc-9lrsh" event={"ID":"7116e0fa-30d9-4a8d-9a11-0d57908f268f","Type":"ContainerDied","Data":"4d4fd8eb0e3f743f4a495e4fd83d8474249ee2494cd13c2d79f583b0d0386669"} Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.261834 4871 scope.go:117] "RemoveContainer" containerID="48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.261847 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cc4f647bc-9lrsh" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.280341 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-config" (OuterVolumeSpecName: "config") pod "7116e0fa-30d9-4a8d-9a11-0d57908f268f" (UID: "7116e0fa-30d9-4a8d-9a11-0d57908f268f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.281192 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7116e0fa-30d9-4a8d-9a11-0d57908f268f" (UID: "7116e0fa-30d9-4a8d-9a11-0d57908f268f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.288480 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7116e0fa-30d9-4a8d-9a11-0d57908f268f" (UID: "7116e0fa-30d9-4a8d-9a11-0d57908f268f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.288551 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "7116e0fa-30d9-4a8d-9a11-0d57908f268f" (UID: "7116e0fa-30d9-4a8d-9a11-0d57908f268f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.301023 4871 scope.go:117] "RemoveContainer" containerID="09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.303190 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.303225 4871 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.303240 4871 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.303253 4871 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.303267 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pmf8\" (UniqueName: \"kubernetes.io/projected/7116e0fa-30d9-4a8d-9a11-0d57908f268f-kube-api-access-5pmf8\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.303280 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-config\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.303293 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7116e0fa-30d9-4a8d-9a11-0d57908f268f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.326501 4871 scope.go:117] "RemoveContainer" containerID="48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a" Oct 07 22:33:11 crc kubenswrapper[4871]: E1007 22:33:11.327109 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a\": container with ID starting with 48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a not found: ID does not exist" containerID="48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.327219 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a"} err="failed to get container status \"48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a\": rpc error: code = NotFound desc = could not find container \"48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a\": container with ID starting with 48f87eff1771f5bc9df526bc14a10b3c83c1e3b77d992d160ea4f824e9e5745a not found: ID does not exist" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.327347 4871 scope.go:117] "RemoveContainer" containerID="09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0" Oct 07 22:33:11 crc kubenswrapper[4871]: E1007 22:33:11.327895 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0\": container with ID starting with 09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0 not found: ID does not exist" containerID="09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.327932 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0"} err="failed to get container status \"09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0\": rpc error: code = NotFound desc = could not find container \"09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0\": container with ID starting with 09427d5d3b1c111faacdacf13d2d654084360f5c6c9b72b0e131ebc202a2a3d0 not found: ID does not exist" Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.614642 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7cc4f647bc-9lrsh"] Oct 07 22:33:11 crc kubenswrapper[4871]: I1007 22:33:11.627489 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7cc4f647bc-9lrsh"] Oct 07 22:33:13 crc kubenswrapper[4871]: I1007 22:33:13.001249 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" path="/var/lib/kubelet/pods/7116e0fa-30d9-4a8d-9a11-0d57908f268f/volumes" Oct 07 22:33:13 crc kubenswrapper[4871]: E1007 22:33:13.582888 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:13 crc kubenswrapper[4871]: E1007 22:33:13.583469 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:13 crc kubenswrapper[4871]: E1007 22:33:13.584042 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:13 crc kubenswrapper[4871]: E1007 22:33:13.584123 4871 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" Oct 07 22:33:13 crc kubenswrapper[4871]: E1007 22:33:13.584947 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:13 crc kubenswrapper[4871]: E1007 22:33:13.587136 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:13 crc kubenswrapper[4871]: E1007 22:33:13.589626 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:13 crc kubenswrapper[4871]: E1007 22:33:13.589695 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" Oct 07 22:33:18 crc kubenswrapper[4871]: E1007 22:33:18.583061 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:18 crc kubenswrapper[4871]: E1007 22:33:18.583687 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:18 crc kubenswrapper[4871]: E1007 22:33:18.584160 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:18 crc kubenswrapper[4871]: E1007 22:33:18.584198 4871 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" Oct 07 22:33:18 crc kubenswrapper[4871]: E1007 22:33:18.585057 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:18 crc kubenswrapper[4871]: E1007 22:33:18.586785 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:18 crc kubenswrapper[4871]: E1007 22:33:18.591876 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:18 crc kubenswrapper[4871]: E1007 22:33:18.591927 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" Oct 07 22:33:23 crc kubenswrapper[4871]: I1007 22:33:23.449321 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jhh4r_57b1b2c5-f33c-4af0-9707-168eae931b2d/ovs-vswitchd/0.log" Oct 07 22:33:23 crc kubenswrapper[4871]: I1007 22:33:23.451067 4871 generic.go:334] "Generic (PLEG): container finished" podID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" exitCode=137 Oct 07 22:33:23 crc kubenswrapper[4871]: I1007 22:33:23.451128 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jhh4r" event={"ID":"57b1b2c5-f33c-4af0-9707-168eae931b2d","Type":"ContainerDied","Data":"e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c"} Oct 07 22:33:23 crc kubenswrapper[4871]: E1007 22:33:23.582580 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:23 crc kubenswrapper[4871]: E1007 22:33:23.583685 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:23 crc kubenswrapper[4871]: E1007 22:33:23.583951 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c is running failed: container process not found" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:23 crc kubenswrapper[4871]: E1007 22:33:23.584438 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c is running failed: container process not found" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:23 crc kubenswrapper[4871]: E1007 22:33:23.584586 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 07 22:33:23 crc kubenswrapper[4871]: E1007 22:33:23.584617 4871 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" Oct 07 22:33:23 crc kubenswrapper[4871]: E1007 22:33:23.585110 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c is running failed: container process not found" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 07 22:33:23 crc kubenswrapper[4871]: E1007 22:33:23.585176 4871 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-jhh4r" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.032521 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jhh4r_57b1b2c5-f33c-4af0-9707-168eae931b2d/ovs-vswitchd/0.log" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.033297 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.119481 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.160612 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57b1b2c5-f33c-4af0-9707-168eae931b2d-scripts\") pod \"57b1b2c5-f33c-4af0-9707-168eae931b2d\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.160669 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-log\") pod \"57b1b2c5-f33c-4af0-9707-168eae931b2d\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.160745 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-combined-ca-bundle\") pod \"825fba74-f028-4445-b041-465ccbbad763\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.160780 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fc4p\" (UniqueName: \"kubernetes.io/projected/57b1b2c5-f33c-4af0-9707-168eae931b2d-kube-api-access-2fc4p\") pod \"57b1b2c5-f33c-4af0-9707-168eae931b2d\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.160838 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg9gt\" (UniqueName: \"kubernetes.io/projected/825fba74-f028-4445-b041-465ccbbad763-kube-api-access-jg9gt\") pod \"825fba74-f028-4445-b041-465ccbbad763\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.160856 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-lib\") pod \"57b1b2c5-f33c-4af0-9707-168eae931b2d\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.160926 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/825fba74-f028-4445-b041-465ccbbad763-etc-machine-id\") pod \"825fba74-f028-4445-b041-465ccbbad763\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.160965 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data\") pod \"825fba74-f028-4445-b041-465ccbbad763\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.160995 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-run\") pod \"57b1b2c5-f33c-4af0-9707-168eae931b2d\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.161044 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-etc-ovs\") pod \"57b1b2c5-f33c-4af0-9707-168eae931b2d\" (UID: \"57b1b2c5-f33c-4af0-9707-168eae931b2d\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.161102 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-scripts\") pod \"825fba74-f028-4445-b041-465ccbbad763\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.161123 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data-custom\") pod \"825fba74-f028-4445-b041-465ccbbad763\" (UID: \"825fba74-f028-4445-b041-465ccbbad763\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.161500 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-lib" (OuterVolumeSpecName: "var-lib") pod "57b1b2c5-f33c-4af0-9707-168eae931b2d" (UID: "57b1b2c5-f33c-4af0-9707-168eae931b2d"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.161597 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-run" (OuterVolumeSpecName: "var-run") pod "57b1b2c5-f33c-4af0-9707-168eae931b2d" (UID: "57b1b2c5-f33c-4af0-9707-168eae931b2d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.161703 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/825fba74-f028-4445-b041-465ccbbad763-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "825fba74-f028-4445-b041-465ccbbad763" (UID: "825fba74-f028-4445-b041-465ccbbad763"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.161742 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-log" (OuterVolumeSpecName: "var-log") pod "57b1b2c5-f33c-4af0-9707-168eae931b2d" (UID: "57b1b2c5-f33c-4af0-9707-168eae931b2d"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.161680 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "57b1b2c5-f33c-4af0-9707-168eae931b2d" (UID: "57b1b2c5-f33c-4af0-9707-168eae931b2d"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.163718 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57b1b2c5-f33c-4af0-9707-168eae931b2d-scripts" (OuterVolumeSpecName: "scripts") pod "57b1b2c5-f33c-4af0-9707-168eae931b2d" (UID: "57b1b2c5-f33c-4af0-9707-168eae931b2d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.168145 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "825fba74-f028-4445-b041-465ccbbad763" (UID: "825fba74-f028-4445-b041-465ccbbad763"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.168719 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b1b2c5-f33c-4af0-9707-168eae931b2d-kube-api-access-2fc4p" (OuterVolumeSpecName: "kube-api-access-2fc4p") pod "57b1b2c5-f33c-4af0-9707-168eae931b2d" (UID: "57b1b2c5-f33c-4af0-9707-168eae931b2d"). InnerVolumeSpecName "kube-api-access-2fc4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.170199 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/825fba74-f028-4445-b041-465ccbbad763-kube-api-access-jg9gt" (OuterVolumeSpecName: "kube-api-access-jg9gt") pod "825fba74-f028-4445-b041-465ccbbad763" (UID: "825fba74-f028-4445-b041-465ccbbad763"). InnerVolumeSpecName "kube-api-access-jg9gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.188020 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-scripts" (OuterVolumeSpecName: "scripts") pod "825fba74-f028-4445-b041-465ccbbad763" (UID: "825fba74-f028-4445-b041-465ccbbad763"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.210897 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "825fba74-f028-4445-b041-465ccbbad763" (UID: "825fba74-f028-4445-b041-465ccbbad763"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262451 4871 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-lib\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262860 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg9gt\" (UniqueName: \"kubernetes.io/projected/825fba74-f028-4445-b041-465ccbbad763-kube-api-access-jg9gt\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262875 4871 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/825fba74-f028-4445-b041-465ccbbad763-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262885 4871 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-run\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262894 4871 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262903 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262913 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262923 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57b1b2c5-f33c-4af0-9707-168eae931b2d-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262930 4871 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/57b1b2c5-f33c-4af0-9707-168eae931b2d-var-log\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262938 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.262946 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fc4p\" (UniqueName: \"kubernetes.io/projected/57b1b2c5-f33c-4af0-9707-168eae931b2d-kube-api-access-2fc4p\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.264298 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.265878 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data" (OuterVolumeSpecName: "config-data") pod "825fba74-f028-4445-b041-465ccbbad763" (UID: "825fba74-f028-4445-b041-465ccbbad763"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.364437 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-lock\") pod \"bb4b36fd-1835-4beb-81fb-0df36301f700\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.364478 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") pod \"bb4b36fd-1835-4beb-81fb-0df36301f700\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.364565 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-cache\") pod \"bb4b36fd-1835-4beb-81fb-0df36301f700\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.364590 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27nsr\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-kube-api-access-27nsr\") pod \"bb4b36fd-1835-4beb-81fb-0df36301f700\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.364663 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"bb4b36fd-1835-4beb-81fb-0df36301f700\" (UID: \"bb4b36fd-1835-4beb-81fb-0df36301f700\") " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.365008 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825fba74-f028-4445-b041-465ccbbad763-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.365896 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-lock" (OuterVolumeSpecName: "lock") pod "bb4b36fd-1835-4beb-81fb-0df36301f700" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.366007 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-cache" (OuterVolumeSpecName: "cache") pod "bb4b36fd-1835-4beb-81fb-0df36301f700" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.367656 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "swift") pod "bb4b36fd-1835-4beb-81fb-0df36301f700" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.370444 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-kube-api-access-27nsr" (OuterVolumeSpecName: "kube-api-access-27nsr") pod "bb4b36fd-1835-4beb-81fb-0df36301f700" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700"). InnerVolumeSpecName "kube-api-access-27nsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.370528 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bb4b36fd-1835-4beb-81fb-0df36301f700" (UID: "bb4b36fd-1835-4beb-81fb-0df36301f700"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.466452 4871 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-lock\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.466526 4871 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.466544 4871 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/bb4b36fd-1835-4beb-81fb-0df36301f700-cache\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.466559 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27nsr\" (UniqueName: \"kubernetes.io/projected/bb4b36fd-1835-4beb-81fb-0df36301f700-kube-api-access-27nsr\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.466644 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.469905 4871 generic.go:334] "Generic (PLEG): container finished" podID="825fba74-f028-4445-b041-465ccbbad763" containerID="ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9" exitCode=137 Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.470098 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.470880 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"825fba74-f028-4445-b041-465ccbbad763","Type":"ContainerDied","Data":"ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9"} Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.470963 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"825fba74-f028-4445-b041-465ccbbad763","Type":"ContainerDied","Data":"00a248eb0776a91ccd453b5b36dc1d36cec9eadd09664079556cc912f23bbbd8"} Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.470996 4871 scope.go:117] "RemoveContainer" containerID="b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.478997 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jhh4r_57b1b2c5-f33c-4af0-9707-168eae931b2d/ovs-vswitchd/0.log" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.480763 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jhh4r" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.480694 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jhh4r" event={"ID":"57b1b2c5-f33c-4af0-9707-168eae931b2d","Type":"ContainerDied","Data":"71c9fab85a529e5e7054bde6eef98e050dedc47387775d33478e2c1e1b5eafc2"} Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.494495 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerID="d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad" exitCode=137 Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.494719 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad"} Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.494778 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"bb4b36fd-1835-4beb-81fb-0df36301f700","Type":"ContainerDied","Data":"67cf4c1b020198e8c306f3bf42d00cf8f2a48766602384e5d3238451532dfb48"} Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.495315 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.495468 4871 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.568609 4871 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.609601 4871 scope.go:117] "RemoveContainer" containerID="ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.620195 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.631358 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.654436 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-jhh4r"] Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.666703 4871 scope.go:117] "RemoveContainer" containerID="b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1" Oct 07 22:33:24 crc kubenswrapper[4871]: E1007 22:33:24.667267 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1\": container with ID starting with b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1 not found: ID does not exist" containerID="b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.667301 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1"} err="failed to get container status \"b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1\": rpc error: code = NotFound desc = could not find container \"b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1\": container with ID starting with b92eb3f9387bd70801fb5f477ddba21b83760ae7b1fa2c398ca78334beb83dc1 not found: ID does not exist" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.667326 4871 scope.go:117] "RemoveContainer" containerID="ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9" Oct 07 22:33:24 crc kubenswrapper[4871]: E1007 22:33:24.667742 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9\": container with ID starting with ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9 not found: ID does not exist" containerID="ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.667780 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9"} err="failed to get container status \"ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9\": rpc error: code = NotFound desc = could not find container \"ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9\": container with ID starting with ee3c734bbda1d50aceec6e6ae5a1bf5baace7620b0f1197921377b6765710bb9 not found: ID does not exist" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.667809 4871 scope.go:117] "RemoveContainer" containerID="e94bc3d77a8f50c202553e3fe28dda3c0297702165742028aed16ce961e8204c" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.669415 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-jhh4r"] Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.675861 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.681425 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.695882 4871 scope.go:117] "RemoveContainer" containerID="2db68696c04f1b0fcc1c3cb7f731d227e8e395c524c2d8943279b5ae68dc711c" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.720652 4871 scope.go:117] "RemoveContainer" containerID="50ba48bfa54cc28d807c8ba1423f3e7544d91a400fe134f0353833893c0e892b" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.748826 4871 scope.go:117] "RemoveContainer" containerID="d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.784255 4871 scope.go:117] "RemoveContainer" containerID="1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.815375 4871 scope.go:117] "RemoveContainer" containerID="5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.847108 4871 scope.go:117] "RemoveContainer" containerID="2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.879326 4871 scope.go:117] "RemoveContainer" containerID="1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.907823 4871 scope.go:117] "RemoveContainer" containerID="8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.937213 4871 scope.go:117] "RemoveContainer" containerID="e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057" Oct 07 22:33:24 crc kubenswrapper[4871]: I1007 22:33:24.960167 4871 scope.go:117] "RemoveContainer" containerID="d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.012683 4871 scope.go:117] "RemoveContainer" containerID="26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.020953 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" path="/var/lib/kubelet/pods/57b1b2c5-f33c-4af0-9707-168eae931b2d/volumes" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.022506 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="825fba74-f028-4445-b041-465ccbbad763" path="/var/lib/kubelet/pods/825fba74-f028-4445-b041-465ccbbad763/volumes" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.024095 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" path="/var/lib/kubelet/pods/bb4b36fd-1835-4beb-81fb-0df36301f700/volumes" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.046951 4871 scope.go:117] "RemoveContainer" containerID="e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.073473 4871 scope.go:117] "RemoveContainer" containerID="a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.107388 4871 scope.go:117] "RemoveContainer" containerID="ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.141390 4871 scope.go:117] "RemoveContainer" containerID="e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.177557 4871 scope.go:117] "RemoveContainer" containerID="35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.214083 4871 scope.go:117] "RemoveContainer" containerID="2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.251608 4871 scope.go:117] "RemoveContainer" containerID="d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.252671 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad\": container with ID starting with d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad not found: ID does not exist" containerID="d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.252746 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad"} err="failed to get container status \"d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad\": rpc error: code = NotFound desc = could not find container \"d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad\": container with ID starting with d909d0512eea8b02a246b35c03e165572e72226b843bcfc1a1dc483e4ddcbcad not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.252818 4871 scope.go:117] "RemoveContainer" containerID="1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.253555 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9\": container with ID starting with 1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9 not found: ID does not exist" containerID="1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.253618 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9"} err="failed to get container status \"1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9\": rpc error: code = NotFound desc = could not find container \"1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9\": container with ID starting with 1b63dfbc1c92748cd8fbf4dfe534bbf0c681f698621b150e685b8fe52f2601f9 not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.253658 4871 scope.go:117] "RemoveContainer" containerID="5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.254259 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d\": container with ID starting with 5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d not found: ID does not exist" containerID="5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.254307 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d"} err="failed to get container status \"5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d\": rpc error: code = NotFound desc = could not find container \"5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d\": container with ID starting with 5891bddd93d251509fc82e0ad0815196183a77f178e12e80f4d6abde7f5e275d not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.254336 4871 scope.go:117] "RemoveContainer" containerID="2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.254761 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778\": container with ID starting with 2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778 not found: ID does not exist" containerID="2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.254820 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778"} err="failed to get container status \"2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778\": rpc error: code = NotFound desc = could not find container \"2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778\": container with ID starting with 2edbb5dcf46541037c72350c418c724304e4e335f51960a57f728352ab5a4778 not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.254850 4871 scope.go:117] "RemoveContainer" containerID="1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.255265 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a\": container with ID starting with 1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a not found: ID does not exist" containerID="1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.255351 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a"} err="failed to get container status \"1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a\": rpc error: code = NotFound desc = could not find container \"1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a\": container with ID starting with 1ea99af89fc9b00a61a0b683be1c29bcb0621ca27dd10f6391c6ca1d03fc059a not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.255380 4871 scope.go:117] "RemoveContainer" containerID="8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.256374 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c\": container with ID starting with 8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c not found: ID does not exist" containerID="8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.256425 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c"} err="failed to get container status \"8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c\": rpc error: code = NotFound desc = could not find container \"8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c\": container with ID starting with 8218459bb3d498fa2f15ae33f7d3e61b8a14a718ace49023b1c7a8d68d8d816c not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.256453 4871 scope.go:117] "RemoveContainer" containerID="e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.257043 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057\": container with ID starting with e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057 not found: ID does not exist" containerID="e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.257083 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057"} err="failed to get container status \"e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057\": rpc error: code = NotFound desc = could not find container \"e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057\": container with ID starting with e983c71665852e0e909a95a9d3d9fadeadac3cdcbe6acfd0623f263d2ba4f057 not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.257110 4871 scope.go:117] "RemoveContainer" containerID="d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.257555 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a\": container with ID starting with d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a not found: ID does not exist" containerID="d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.257621 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a"} err="failed to get container status \"d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a\": rpc error: code = NotFound desc = could not find container \"d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a\": container with ID starting with d9fee2bba3f233515ec831f5d7fbe498b8c8a6e7a252ee935a4df973e718383a not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.257673 4871 scope.go:117] "RemoveContainer" containerID="26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.258134 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c\": container with ID starting with 26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c not found: ID does not exist" containerID="26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.258179 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c"} err="failed to get container status \"26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c\": rpc error: code = NotFound desc = could not find container \"26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c\": container with ID starting with 26cf1d2ac396d79ffbc5391e86b0a821d3120e9469497dfa3901c3ec146fa60c not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.258208 4871 scope.go:117] "RemoveContainer" containerID="e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.258660 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315\": container with ID starting with e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315 not found: ID does not exist" containerID="e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.258705 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315"} err="failed to get container status \"e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315\": rpc error: code = NotFound desc = could not find container \"e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315\": container with ID starting with e8affd9de1015f26ca069bc5b16a3a35d710ff379598af84f8709726a0d07315 not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.258739 4871 scope.go:117] "RemoveContainer" containerID="a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.259141 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8\": container with ID starting with a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8 not found: ID does not exist" containerID="a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.259195 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8"} err="failed to get container status \"a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8\": rpc error: code = NotFound desc = could not find container \"a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8\": container with ID starting with a29ce9d391c3a436c72e8d294787ac03ab8f1df3b1ffb0b9e255d9dbc7495ec8 not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.259231 4871 scope.go:117] "RemoveContainer" containerID="ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.260067 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae\": container with ID starting with ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae not found: ID does not exist" containerID="ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.260114 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae"} err="failed to get container status \"ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae\": rpc error: code = NotFound desc = could not find container \"ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae\": container with ID starting with ed3a16b4a54fc54ef0de1eaa8443289f4018f66919146e860b8dcffcb48b53ae not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.260146 4871 scope.go:117] "RemoveContainer" containerID="e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.260632 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c\": container with ID starting with e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c not found: ID does not exist" containerID="e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.260679 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c"} err="failed to get container status \"e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c\": rpc error: code = NotFound desc = could not find container \"e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c\": container with ID starting with e0314055e3947c2f5ad3fa342638f4e10e6a64fcfd39991116f37a6deba3940c not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.260706 4871 scope.go:117] "RemoveContainer" containerID="35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.261430 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89\": container with ID starting with 35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89 not found: ID does not exist" containerID="35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.261540 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89"} err="failed to get container status \"35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89\": rpc error: code = NotFound desc = could not find container \"35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89\": container with ID starting with 35f2d880581ca125557182b781693ab30cce5ef9a53cfa51bcaf6a1490c25e89 not found: ID does not exist" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.261584 4871 scope.go:117] "RemoveContainer" containerID="2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0" Oct 07 22:33:25 crc kubenswrapper[4871]: E1007 22:33:25.262493 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0\": container with ID starting with 2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0 not found: ID does not exist" containerID="2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0" Oct 07 22:33:25 crc kubenswrapper[4871]: I1007 22:33:25.262584 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0"} err="failed to get container status \"2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0\": rpc error: code = NotFound desc = could not find container \"2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0\": container with ID starting with 2617f51b1cc62b97ec6afd5c91452927bf60f96aaf13cd327290dfac2486f6c0 not found: ID does not exist" Oct 07 22:33:27 crc kubenswrapper[4871]: I1007 22:33:27.828627 4871 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod645a6027-7304-4fdc-aaf4-ffd433bec275"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod645a6027-7304-4fdc-aaf4-ffd433bec275] : Timed out while waiting for systemd to remove kubepods-besteffort-pod645a6027_7304_4fdc_aaf4_ffd433bec275.slice" Oct 07 22:33:27 crc kubenswrapper[4871]: E1007 22:33:27.829190 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod645a6027-7304-4fdc-aaf4-ffd433bec275] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod645a6027-7304-4fdc-aaf4-ffd433bec275] : Timed out while waiting for systemd to remove kubepods-besteffort-pod645a6027_7304_4fdc_aaf4_ffd433bec275.slice" pod="openstack/swift-proxy-7748587bfc-lxnz6" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" Oct 07 22:33:28 crc kubenswrapper[4871]: I1007 22:33:28.569352 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7748587bfc-lxnz6" Oct 07 22:33:28 crc kubenswrapper[4871]: I1007 22:33:28.703583 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7748587bfc-lxnz6"] Oct 07 22:33:28 crc kubenswrapper[4871]: I1007 22:33:28.715261 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-7748587bfc-lxnz6"] Oct 07 22:33:28 crc kubenswrapper[4871]: I1007 22:33:28.964738 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapid46f-account-delete-8z75z" Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.007218 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="645a6027-7304-4fdc-aaf4-ffd433bec275" path="/var/lib/kubelet/pods/645a6027-7304-4fdc-aaf4-ffd433bec275/volumes" Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.085699 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9l76\" (UniqueName: \"kubernetes.io/projected/ca98e71f-5147-4bb7-bd76-e4bb11cfcf42-kube-api-access-q9l76\") pod \"ca98e71f-5147-4bb7-bd76-e4bb11cfcf42\" (UID: \"ca98e71f-5147-4bb7-bd76-e4bb11cfcf42\") " Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.091201 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca98e71f-5147-4bb7-bd76-e4bb11cfcf42-kube-api-access-q9l76" (OuterVolumeSpecName: "kube-api-access-q9l76") pod "ca98e71f-5147-4bb7-bd76-e4bb11cfcf42" (UID: "ca98e71f-5147-4bb7-bd76-e4bb11cfcf42"). InnerVolumeSpecName "kube-api-access-q9l76". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.188020 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9l76\" (UniqueName: \"kubernetes.io/projected/ca98e71f-5147-4bb7-bd76-e4bb11cfcf42-kube-api-access-q9l76\") on node \"crc\" DevicePath \"\"" Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.584612 4871 generic.go:334] "Generic (PLEG): container finished" podID="ca98e71f-5147-4bb7-bd76-e4bb11cfcf42" containerID="655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda" exitCode=137 Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.584678 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapid46f-account-delete-8z75z" event={"ID":"ca98e71f-5147-4bb7-bd76-e4bb11cfcf42","Type":"ContainerDied","Data":"655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda"} Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.584696 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapid46f-account-delete-8z75z" Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.584721 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapid46f-account-delete-8z75z" event={"ID":"ca98e71f-5147-4bb7-bd76-e4bb11cfcf42","Type":"ContainerDied","Data":"58de1ef62139407f0ad019c635b6f4e1d1c507b22628865fb44d4e6cb0dbf48f"} Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.584767 4871 scope.go:117] "RemoveContainer" containerID="655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda" Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.618090 4871 scope.go:117] "RemoveContainer" containerID="655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda" Oct 07 22:33:29 crc kubenswrapper[4871]: E1007 22:33:29.618655 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda\": container with ID starting with 655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda not found: ID does not exist" containerID="655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda" Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.618733 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda"} err="failed to get container status \"655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda\": rpc error: code = NotFound desc = could not find container \"655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda\": container with ID starting with 655e2d6cbfbe6e5d99cab35a9f4d9d4beb443ccf7d94c7696051cc4a2236bcda not found: ID does not exist" Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.643461 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapid46f-account-delete-8z75z"] Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.654734 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapid46f-account-delete-8z75z"] Oct 07 22:33:29 crc kubenswrapper[4871]: I1007 22:33:29.791573 4871 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod878958ac-64b8-434b-88ee-f23d3c486765"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod878958ac-64b8-434b-88ee-f23d3c486765] : Timed out while waiting for systemd to remove kubepods-besteffort-pod878958ac_64b8_434b_88ee_f23d3c486765.slice" Oct 07 22:33:29 crc kubenswrapper[4871]: E1007 22:33:29.791661 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod878958ac-64b8-434b-88ee-f23d3c486765] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod878958ac-64b8-434b-88ee-f23d3c486765] : Timed out while waiting for systemd to remove kubepods-besteffort-pod878958ac_64b8_434b_88ee_f23d3c486765.slice" pod="openstack/placement-7567f46b5d-kb4h9" podUID="878958ac-64b8-434b-88ee-f23d3c486765" Oct 07 22:33:30 crc kubenswrapper[4871]: I1007 22:33:30.597333 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7567f46b5d-kb4h9" Oct 07 22:33:30 crc kubenswrapper[4871]: I1007 22:33:30.632663 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7567f46b5d-kb4h9"] Oct 07 22:33:30 crc kubenswrapper[4871]: I1007 22:33:30.642838 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7567f46b5d-kb4h9"] Oct 07 22:33:30 crc kubenswrapper[4871]: I1007 22:33:30.999179 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="878958ac-64b8-434b-88ee-f23d3c486765" path="/var/lib/kubelet/pods/878958ac-64b8-434b-88ee-f23d3c486765/volumes" Oct 07 22:33:31 crc kubenswrapper[4871]: I1007 22:33:31.000499 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca98e71f-5147-4bb7-bd76-e4bb11cfcf42" path="/var/lib/kubelet/pods/ca98e71f-5147-4bb7-bd76-e4bb11cfcf42/volumes" Oct 07 22:33:35 crc kubenswrapper[4871]: I1007 22:33:35.512543 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:33:35 crc kubenswrapper[4871]: I1007 22:33:35.512891 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:33:50 crc kubenswrapper[4871]: I1007 22:33:50.630766 4871 scope.go:117] "RemoveContainer" containerID="cd6faea6c3b696458cd8465298585cf7714badd7a13ee982186f92932560d733" Oct 07 22:34:05 crc kubenswrapper[4871]: I1007 22:34:05.512303 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:34:05 crc kubenswrapper[4871]: I1007 22:34:05.513120 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:34:05 crc kubenswrapper[4871]: I1007 22:34:05.513212 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:34:05 crc kubenswrapper[4871]: I1007 22:34:05.514519 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:34:05 crc kubenswrapper[4871]: I1007 22:34:05.514635 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" gracePeriod=600 Oct 07 22:34:05 crc kubenswrapper[4871]: E1007 22:34:05.649777 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:34:06 crc kubenswrapper[4871]: I1007 22:34:06.055170 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" exitCode=0 Oct 07 22:34:06 crc kubenswrapper[4871]: I1007 22:34:06.055265 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6"} Oct 07 22:34:06 crc kubenswrapper[4871]: I1007 22:34:06.055352 4871 scope.go:117] "RemoveContainer" containerID="a21e914393b384baa70031a146416b2ffe44375e4ba7b43b6aad9a92b38797c6" Oct 07 22:34:06 crc kubenswrapper[4871]: I1007 22:34:06.056367 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:34:06 crc kubenswrapper[4871]: E1007 22:34:06.056893 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:34:20 crc kubenswrapper[4871]: I1007 22:34:20.983358 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:34:20 crc kubenswrapper[4871]: E1007 22:34:20.984607 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:34:31 crc kubenswrapper[4871]: I1007 22:34:31.982484 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:34:31 crc kubenswrapper[4871]: E1007 22:34:31.983626 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:34:42 crc kubenswrapper[4871]: I1007 22:34:42.982237 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:34:42 crc kubenswrapper[4871]: E1007 22:34:42.983027 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.308829 4871 scope.go:117] "RemoveContainer" containerID="d0224dd10a0847f3e3b8e29d449dcabf476a8b5f0e8ee506885df65b41d807f8" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.345189 4871 scope.go:117] "RemoveContainer" containerID="7da0fc02050c5905620a8024d8e1e85e13be4ceca818e33fb1a1a757f43a3f91" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.373019 4871 scope.go:117] "RemoveContainer" containerID="4a3cd78bb6c3e29ad022309e0781d2d390b88fc52213fdb0aae6d15cca708568" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.421421 4871 scope.go:117] "RemoveContainer" containerID="dbb6d5909ff3ea7e58ebe374f5b565ded04d14ec719380765bf7aae3bcbbb674" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.453067 4871 scope.go:117] "RemoveContainer" containerID="302e959638c8acd6469fc15e7a7a5fbe4511eec6e971d1b72ea3c8e26f4fea68" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.478448 4871 scope.go:117] "RemoveContainer" containerID="24949ae5014b016a913830d4eb1e8127bec0c97db6d4be6265a57b25735e1294" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.502633 4871 scope.go:117] "RemoveContainer" containerID="c977ba81193a781589b8de9336bdffb902f2bc69bde7a885590c026fb068065b" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.540484 4871 scope.go:117] "RemoveContainer" containerID="f560fa449772ee68ef8cd0288a01316c4f2453ff2605948fb794728f08420983" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.570049 4871 scope.go:117] "RemoveContainer" containerID="52ca931b6e665fac4e12d71a04df651d6da1645c3ce0598832d89f347ee1e924" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.615938 4871 scope.go:117] "RemoveContainer" containerID="ee6089abc123f34af422521b137cc52ea7e71cd2d1835d9222589cfa258c4a2b" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.642205 4871 scope.go:117] "RemoveContainer" containerID="4a7a0515437b353511dce4852c471123dc48a9388970a160c88a5a0ae9a250d1" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.686536 4871 scope.go:117] "RemoveContainer" containerID="3cafaed81f8618ad2d607a8694a1aa368e23d06c9cfb94bab5130c43974b6205" Oct 07 22:34:51 crc kubenswrapper[4871]: I1007 22:34:51.717802 4871 scope.go:117] "RemoveContainer" containerID="587e087735b583679a4e0927c1e92977a3a6e911ca84ae72de2b03c4a34930de" Oct 07 22:34:57 crc kubenswrapper[4871]: I1007 22:34:57.982109 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:34:57 crc kubenswrapper[4871]: E1007 22:34:57.982841 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:35:12 crc kubenswrapper[4871]: I1007 22:35:12.983958 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:35:12 crc kubenswrapper[4871]: E1007 22:35:12.985622 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:35:24 crc kubenswrapper[4871]: I1007 22:35:24.983180 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:35:24 crc kubenswrapper[4871]: E1007 22:35:24.983939 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:35:37 crc kubenswrapper[4871]: I1007 22:35:37.983190 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:35:37 crc kubenswrapper[4871]: E1007 22:35:37.984206 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:35:50 crc kubenswrapper[4871]: I1007 22:35:50.982470 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:35:50 crc kubenswrapper[4871]: E1007 22:35:50.983543 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.082007 4871 scope.go:117] "RemoveContainer" containerID="6df113a7ac9f9d5f536d35f13f7fad51d2a38b5a2ccec4a5d4e7b944182cadb4" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.122411 4871 scope.go:117] "RemoveContainer" containerID="15a075e9bb05d488c246bfc9ab9f1e3bce81bc15386a51cca52b6ab00f6b66e8" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.179185 4871 scope.go:117] "RemoveContainer" containerID="3ce17290b46def925cf8b9b69d4142ed8a8f644537f24587c21309531f91602e" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.206139 4871 scope.go:117] "RemoveContainer" containerID="9ae44c9c6b44d826a4bf86d7e4c7289e841ebde0a866bf0d9aa672e7b5e7c0bd" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.262045 4871 scope.go:117] "RemoveContainer" containerID="650b1a6f51292f66150c255a0c37fc55a4deb4ccee8e73d591fe49e09f2a0754" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.286924 4871 scope.go:117] "RemoveContainer" containerID="e6f0d276a1846b36e45c6f4882a0d1c7b6023b4bc31e6f161d4fcc4d02c212f3" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.314248 4871 scope.go:117] "RemoveContainer" containerID="673a2fb678380df6b3c6dd2cb9e872c1ff636cea1ee56f22ae5bf5386ab04451" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.375231 4871 scope.go:117] "RemoveContainer" containerID="a4fb5e314a418ad3f5df08f22eefe7f38b31a60269157a06a8f3094f45ee90eb" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.405604 4871 scope.go:117] "RemoveContainer" containerID="8ed9776a16b485c2c2642bcabf5d61e86921b8b52f37c2610b01e764373ee95c" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.444662 4871 scope.go:117] "RemoveContainer" containerID="cf553428752c5df543a55de342dcca65031384272035f4e1dedff261e60e8752" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.480934 4871 scope.go:117] "RemoveContainer" containerID="db14634af4bd8ffd95be4d16f4dbd649bbf43ca5ba952ff1a56f66dcbab97868" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.510220 4871 scope.go:117] "RemoveContainer" containerID="e9b84dbda4c0fa3402e53853a13275e7a4791285459387786a2900afc2dbb3a7" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.548147 4871 scope.go:117] "RemoveContainer" containerID="fa8c238c4e4dd8b2b2f155c5c018100d3ed579f7df72b59d5814579618e15651" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.588255 4871 scope.go:117] "RemoveContainer" containerID="5b46a03277275d9c2cc38bd8814c0a9e068549a5c9ce5f29fba665667d373c7a" Oct 07 22:35:52 crc kubenswrapper[4871]: I1007 22:35:52.648735 4871 scope.go:117] "RemoveContainer" containerID="0c8b1c9fd16dade7ae8c609c61cead66d6386df85d637f9397df20edbb885c4a" Oct 07 22:36:05 crc kubenswrapper[4871]: I1007 22:36:05.982388 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:36:05 crc kubenswrapper[4871]: E1007 22:36:05.983414 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:36:20 crc kubenswrapper[4871]: I1007 22:36:20.982553 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:36:20 crc kubenswrapper[4871]: E1007 22:36:20.983386 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:36:35 crc kubenswrapper[4871]: I1007 22:36:35.982853 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:36:35 crc kubenswrapper[4871]: E1007 22:36:35.984117 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:36:49 crc kubenswrapper[4871]: I1007 22:36:49.982545 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:36:49 crc kubenswrapper[4871]: E1007 22:36:49.983866 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:36:52 crc kubenswrapper[4871]: I1007 22:36:52.885732 4871 scope.go:117] "RemoveContainer" containerID="ee44fec5c9a0421e422d1f7b00f0a817126ddfb1c416825eb1642a57c8becf28" Oct 07 22:36:52 crc kubenswrapper[4871]: I1007 22:36:52.915240 4871 scope.go:117] "RemoveContainer" containerID="3172ad3a17b669b3f933bc0df5ba7c55bf17064fd77c9124684cfa839ca3bd73" Oct 07 22:36:52 crc kubenswrapper[4871]: I1007 22:36:52.962964 4871 scope.go:117] "RemoveContainer" containerID="a6bcbbe481eb19cf659465c0d4fcb0fa01acc7646b80f2eaea5e2ac2187ebf78" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.012096 4871 scope.go:117] "RemoveContainer" containerID="24dbb24a65b2912e92cb5795b73a40b47b50468929436754aa78a71953f92649" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.034561 4871 scope.go:117] "RemoveContainer" containerID="91eb51510e831d42f41a70585ae844bb55e88150c93eae1571a50da925c20bf3" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.076379 4871 scope.go:117] "RemoveContainer" containerID="8348036e75afc1a3c2a675f95267987dea996bddef522f920d21ab3ccc437dfd" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.097678 4871 scope.go:117] "RemoveContainer" containerID="303e52fe41cb9eb2c55a65b8e6fb2c26f6e89c632f0d230eb1f5ba71eb482fee" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.129071 4871 scope.go:117] "RemoveContainer" containerID="0b297eaadbd30609ad30865b36ac820cfdedfdd27b2cf2f2a38dea46f5c786d9" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.153930 4871 scope.go:117] "RemoveContainer" containerID="d4a49dd5137df3e1d29d45f30f2df5cef25e935b8d3ca458065782634e5b1f86" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.175474 4871 scope.go:117] "RemoveContainer" containerID="1c378db964db6ca85fc9a95ee1643c1a2799b62adc25117e536465d06965954e" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.249468 4871 scope.go:117] "RemoveContainer" containerID="4bd3d3abc247db1b0ff7aa3bc5abe1e19e8cfebb175b85960154e77696a3ac5d" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.273168 4871 scope.go:117] "RemoveContainer" containerID="286c54211026600e80566fbac1ca116b1dff1c146ef3a9a8fb05b27ac63ebc82" Oct 07 22:36:53 crc kubenswrapper[4871]: I1007 22:36:53.303642 4871 scope.go:117] "RemoveContainer" containerID="5e73453b1f98ecb0f88fa594bdc169a23d7bfd230c5e5e2791d2410183915462" Oct 07 22:37:01 crc kubenswrapper[4871]: I1007 22:37:01.982535 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:37:01 crc kubenswrapper[4871]: E1007 22:37:01.983299 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:37:15 crc kubenswrapper[4871]: I1007 22:37:15.982886 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:37:15 crc kubenswrapper[4871]: E1007 22:37:15.984155 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:37:29 crc kubenswrapper[4871]: I1007 22:37:29.982346 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:37:29 crc kubenswrapper[4871]: E1007 22:37:29.983478 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:37:42 crc kubenswrapper[4871]: I1007 22:37:42.984031 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:37:42 crc kubenswrapper[4871]: E1007 22:37:42.985295 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:37:53 crc kubenswrapper[4871]: I1007 22:37:53.519261 4871 scope.go:117] "RemoveContainer" containerID="c09d7323fe49b6e9493519e2e016807782896bb28fb648e44c167ed8fa4b538a" Oct 07 22:37:53 crc kubenswrapper[4871]: I1007 22:37:53.564580 4871 scope.go:117] "RemoveContainer" containerID="ac7926263bebb46db052a48e0d59ffb573ae16e631980ef5e58f7119d8da1a49" Oct 07 22:37:53 crc kubenswrapper[4871]: I1007 22:37:53.604996 4871 scope.go:117] "RemoveContainer" containerID="525b8530afe70e73d50536a5be4989228b6800d6312af9fbc65f81c15ad00346" Oct 07 22:37:53 crc kubenswrapper[4871]: I1007 22:37:53.666321 4871 scope.go:117] "RemoveContainer" containerID="45785f0d795f8ab376fa609b6da3c877ac64fc1ed8340a47e9be2e197d469507" Oct 07 22:37:53 crc kubenswrapper[4871]: I1007 22:37:53.721691 4871 scope.go:117] "RemoveContainer" containerID="2ff1e6c6e2b4bf5a4e4fe91286071d2d9c47cf4d7640e8db226f47e4bb0157a4" Oct 07 22:37:57 crc kubenswrapper[4871]: I1007 22:37:57.982491 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:37:57 crc kubenswrapper[4871]: E1007 22:37:57.983291 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:38:12 crc kubenswrapper[4871]: I1007 22:38:12.983178 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:38:12 crc kubenswrapper[4871]: E1007 22:38:12.984469 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:38:27 crc kubenswrapper[4871]: I1007 22:38:27.982262 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:38:27 crc kubenswrapper[4871]: E1007 22:38:27.983554 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:38:38 crc kubenswrapper[4871]: I1007 22:38:38.985580 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:38:38 crc kubenswrapper[4871]: E1007 22:38:38.988634 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:38:50 crc kubenswrapper[4871]: I1007 22:38:50.982765 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:38:50 crc kubenswrapper[4871]: E1007 22:38:50.983952 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:38:53 crc kubenswrapper[4871]: I1007 22:38:53.883103 4871 scope.go:117] "RemoveContainer" containerID="893dcbed9540fca5394cb10bce40d68accf0e792ccb53a0bca71eeb8cf8337c3" Oct 07 22:38:53 crc kubenswrapper[4871]: I1007 22:38:53.914692 4871 scope.go:117] "RemoveContainer" containerID="41dcab2800cab633133f6b01f760c087220835afec33308c322b94a96c4c33d9" Oct 07 22:38:53 crc kubenswrapper[4871]: I1007 22:38:53.992577 4871 scope.go:117] "RemoveContainer" containerID="2bf5bc8c358110db521839c0ee5aeb6e45d6fb3a5fe434cba3a0bfc2e3f361c7" Oct 07 22:38:54 crc kubenswrapper[4871]: I1007 22:38:54.016451 4871 scope.go:117] "RemoveContainer" containerID="83a8879c5c8652cd944b767603633f0434212ef09dde9ebec772c3c1fe6093a1" Oct 07 22:38:54 crc kubenswrapper[4871]: I1007 22:38:54.040150 4871 scope.go:117] "RemoveContainer" containerID="0e1f568994d1e614f92908e5ccc4ef98ce4c080d84575da81deff6e86b817516" Oct 07 22:39:03 crc kubenswrapper[4871]: I1007 22:39:03.982860 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:39:03 crc kubenswrapper[4871]: E1007 22:39:03.984015 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:39:16 crc kubenswrapper[4871]: I1007 22:39:16.989618 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:39:17 crc kubenswrapper[4871]: I1007 22:39:17.593421 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"ecf3f7625fa1b5db1c3a7b56d23927b826eefe23c4c9f7a906222c77a0c97263"} Oct 07 22:39:54 crc kubenswrapper[4871]: I1007 22:39:54.128739 4871 scope.go:117] "RemoveContainer" containerID="b1ba6099d51e0bd9f15eb3e0c286b7e8227ca05e044b331b744976e0d99a05b9" Oct 07 22:39:54 crc kubenswrapper[4871]: I1007 22:39:54.163187 4871 scope.go:117] "RemoveContainer" containerID="336725751a77fc6966bb612d7aba14caa7fe6292b10886ecbb69587c4cead039" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.466784 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sfvz2"] Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.467840 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-reaper" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.467855 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-reaper" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.467875 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-updater" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.467884 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-updater" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.467898 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerName="neutron-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.467907 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerName="neutron-api" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.467925 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd77ac8e-36b6-4e16-869a-3ce6bff223ac" containerName="setup-container" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.467933 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd77ac8e-36b6-4e16-869a-3ce6bff223ac" containerName="setup-container" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.467948 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="878958ac-64b8-434b-88ee-f23d3c486765" containerName="placement-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.467956 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="878958ac-64b8-434b-88ee-f23d3c486765" containerName="placement-log" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.467970 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="sg-core" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.467979 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="sg-core" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.467990 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd77ac8e-36b6-4e16-869a-3ce6bff223ac" containerName="rabbitmq" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.467998 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd77ac8e-36b6-4e16-869a-3ce6bff223ac" containerName="rabbitmq" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468011 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e61b0b6-83cc-48b0-9f3d-b9b607faa830" containerName="nova-cell0-conductor-conductor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468018 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e61b0b6-83cc-48b0-9f3d-b9b607faa830" containerName="nova-cell0-conductor-conductor" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468027 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-expirer" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468035 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-expirer" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468049 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829bd903-9e3e-4378-9c56-6a6375d26422" containerName="kube-state-metrics" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468057 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="829bd903-9e3e-4378-9c56-6a6375d26422" containerName="kube-state-metrics" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468068 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server-init" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468076 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server-init" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468088 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17ba7e1-deb0-4a31-8874-ee1e6686a3f8" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468095 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17ba7e1-deb0-4a31-8874-ee1e6686a3f8" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468108 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-auditor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468115 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-auditor" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468126 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-auditor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468134 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-auditor" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468146 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468154 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468165 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468173 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-log" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468183 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="ceilometer-central-agent" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468191 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="ceilometer-central-agent" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468201 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-metadata" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468209 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-metadata" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468219 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6af0bcb3-3803-4174-af06-c6841610c62a" containerName="keystone-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468227 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6af0bcb3-3803-4174-af06-c6841610c62a" containerName="keystone-api" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468235 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="ceilometer-notification-agent" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468245 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="ceilometer-notification-agent" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468255 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="ovn-northd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468263 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="ovn-northd" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468278 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" containerName="setup-container" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468286 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" containerName="setup-container" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468295 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-auditor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468302 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-auditor" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468315 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerName="glance-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468323 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerName="glance-log" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468332 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca98e71f-5147-4bb7-bd76-e4bb11cfcf42" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468340 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca98e71f-5147-4bb7-bd76-e4bb11cfcf42" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468353 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7d1def-9358-44f0-9245-e333fb60bac4" containerName="nova-scheduler-scheduler" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468361 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7d1def-9358-44f0-9245-e333fb60bac4" containerName="nova-scheduler-scheduler" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468370 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825fba74-f028-4445-b041-465ccbbad763" containerName="cinder-scheduler" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468377 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="825fba74-f028-4445-b041-465ccbbad763" containerName="cinder-scheduler" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468385 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="proxy-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468393 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="proxy-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468407 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerName="neutron-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468415 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerName="neutron-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468423 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468431 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-api" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468443 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22210cb6-0be2-4f3d-9e04-d36274391d54" containerName="ovn-controller" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468451 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="22210cb6-0be2-4f3d-9e04-d36274391d54" containerName="ovn-controller" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468466 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468473 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468507 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-server" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468515 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-server" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468529 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerName="glance-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468537 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerName="glance-log" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468550 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f602e068-0d3e-4f32-91ca-3ec36c965a7c" containerName="memcached" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468558 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f602e068-0d3e-4f32-91ca-3ec36c965a7c" containerName="memcached" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468568 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-updater" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468575 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-updater" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468583 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerName="glance-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468591 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerName="glance-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468601 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" containerName="cinder-api-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468609 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" containerName="cinder-api-log" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468622 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-server" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468630 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-server" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468638 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerName="barbican-worker" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468646 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerName="barbican-worker" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468658 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-server" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468666 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-server" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468678 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468688 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468698 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825fba74-f028-4445-b041-465ccbbad763" containerName="probe" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468706 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="825fba74-f028-4445-b041-465ccbbad763" containerName="probe" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468718 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8724dd55-b509-4056-b324-531ec57c3a7e" containerName="barbican-keystone-listener" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468727 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8724dd55-b509-4056-b324-531ec57c3a7e" containerName="barbican-keystone-listener" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468738 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468745 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468756 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" containerName="cinder-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468765 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" containerName="cinder-api" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468775 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="361d6d85-f8c7-4643-95c4-29741f91a179" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468782 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="361d6d85-f8c7-4643-95c4-29741f91a179" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468816 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-replicator" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468824 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-replicator" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468835 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="openstack-network-exporter" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468843 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="openstack-network-exporter" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468854 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="878958ac-64b8-434b-88ee-f23d3c486765" containerName="placement-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468861 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="878958ac-64b8-434b-88ee-f23d3c486765" containerName="placement-api" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468875 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd68bd87-c8ef-4214-a9b2-299efdcb956e" containerName="nova-cell1-conductor-conductor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468883 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd68bd87-c8ef-4214-a9b2-299efdcb956e" containerName="nova-cell1-conductor-conductor" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468896 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e4de98-7d7d-43af-a76b-abf1774c620d" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468904 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e4de98-7d7d-43af-a76b-abf1774c620d" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468920 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-replicator" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468927 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-replicator" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468942 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54142d12-b692-4acd-83f7-e3c029b69d3a" containerName="galera" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468949 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="54142d12-b692-4acd-83f7-e3c029b69d3a" containerName="galera" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468961 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="rsync" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468968 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="rsync" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.468981 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8724dd55-b509-4056-b324-531ec57c3a7e" containerName="barbican-keystone-listener-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.468989 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8724dd55-b509-4056-b324-531ec57c3a7e" containerName="barbican-keystone-listener-log" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.469000 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerName="barbican-worker-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469008 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerName="barbican-worker-log" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.469020 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="swift-recon-cron" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469028 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="swift-recon-cron" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.469036 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469044 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-log" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.469057 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerName="glance-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469064 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerName="glance-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.469074 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-replicator" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469081 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-replicator" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.469095 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54142d12-b692-4acd-83f7-e3c029b69d3a" containerName="mysql-bootstrap" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469103 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="54142d12-b692-4acd-83f7-e3c029b69d3a" containerName="mysql-bootstrap" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.469116 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" containerName="rabbitmq" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469123 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" containerName="rabbitmq" Oct 07 22:40:57 crc kubenswrapper[4871]: E1007 22:40:57.469136 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469144 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469309 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerName="barbican-worker" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469320 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="proxy-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469331 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469344 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="rsync" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469358 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469369 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-auditor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469381 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-replicator" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469390 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-replicator" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469400 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="825fba74-f028-4445-b041-465ccbbad763" containerName="cinder-scheduler" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469408 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="23b48c5c-8fd6-4578-bc71-3f7b47860649" containerName="nova-metadata-metadata" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469419 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="ceilometer-central-agent" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469428 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-updater" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469437 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469446 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e61b0b6-83cc-48b0-9f3d-b9b607faa830" containerName="nova-cell0-conductor-conductor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469456 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-server" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469465 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="825fba74-f028-4445-b041-465ccbbad763" containerName="probe" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469478 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b17ba7e1-deb0-4a31-8874-ee1e6686a3f8" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469488 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="878958ac-64b8-434b-88ee-f23d3c486765" containerName="placement-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469498 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-updater" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469510 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="8724dd55-b509-4056-b324-531ec57c3a7e" containerName="barbican-keystone-listener" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469524 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="ovn-northd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469533 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" containerName="cinder-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469544 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e2b95b3-dbe6-44cb-8f76-0adbf3a83e10" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469555 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-auditor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469564 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="swift-recon-cron" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469572 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e4de98-7d7d-43af-a76b-abf1774c620d" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469583 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7d1def-9358-44f0-9245-e333fb60bac4" containerName="nova-scheduler-scheduler" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469594 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-server" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469607 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="54142d12-b692-4acd-83f7-e3c029b69d3a" containerName="galera" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469618 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-reaper" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469625 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-expirer" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469638 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerName="glance-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469649 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="sg-core" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469663 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="account-replicator" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469672 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerName="glance-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469680 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="object-auditor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469690 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerName="neutron-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469704 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd77ac8e-36b6-4e16-869a-3ce6bff223ac" containerName="rabbitmq" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469712 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4b36fd-1835-4beb-81fb-0df36301f700" containerName="container-server" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469722 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="361d6d85-f8c7-4643-95c4-29741f91a179" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469730 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="22210cb6-0be2-4f3d-9e04-d36274391d54" containerName="ovn-controller" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469743 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovs-vswitchd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469753 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d299e74-bffe-4773-a33a-d615606ecc85" containerName="cinder-api-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469765 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e71edefc-c645-4e74-8077-42df9e751ce0" containerName="ceilometer-notification-agent" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469773 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="8724dd55-b509-4056-b324-531ec57c3a7e" containerName="barbican-keystone-listener-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469785 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9973055-9eec-4599-9d2b-ccef87f6a481" containerName="barbican-worker-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469846 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="c750808e-de6d-442b-82ed-d3498b036d91" containerName="barbican-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469857 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7116e0fa-30d9-4a8d-9a11-0d57908f268f" containerName="neutron-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469866 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ac984d-67d5-4bd1-b3c2-ef99976a8708" containerName="glance-httpd" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469875 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d02631b4-f3a6-48d5-a43e-c90a48622ec2" containerName="glance-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469883 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="829bd903-9e3e-4378-9c56-6a6375d26422" containerName="kube-state-metrics" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469894 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd68bd87-c8ef-4214-a9b2-299efdcb956e" containerName="nova-cell1-conductor-conductor" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469904 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a30eed1-ba36-43aa-9c02-3e4ebe4e02cb" containerName="rabbitmq" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469912 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="878958ac-64b8-434b-88ee-f23d3c486765" containerName="placement-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469922 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="68dde0ec-a90a-4b25-b329-ec645d32671f" containerName="nova-api-log" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469936 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca98e71f-5147-4bb7-bd76-e4bb11cfcf42" containerName="mariadb-account-delete" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469946 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f602e068-0d3e-4f32-91ca-3ec36c965a7c" containerName="memcached" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469955 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb7d0be2-8b44-44ed-948d-11e1932d27b4" containerName="openstack-network-exporter" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469966 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="6af0bcb3-3803-4174-af06-c6841610c62a" containerName="keystone-api" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.469975 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b1b2c5-f33c-4af0-9707-168eae931b2d" containerName="ovsdb-server" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.471266 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.497744 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sfvz2"] Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.593106 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-catalog-content\") pod \"redhat-operators-sfvz2\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.593191 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75r4x\" (UniqueName: \"kubernetes.io/projected/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-kube-api-access-75r4x\") pod \"redhat-operators-sfvz2\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.593243 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-utilities\") pod \"redhat-operators-sfvz2\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.695074 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75r4x\" (UniqueName: \"kubernetes.io/projected/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-kube-api-access-75r4x\") pod \"redhat-operators-sfvz2\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.695151 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-utilities\") pod \"redhat-operators-sfvz2\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.695202 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-catalog-content\") pod \"redhat-operators-sfvz2\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.695640 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-utilities\") pod \"redhat-operators-sfvz2\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.695715 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-catalog-content\") pod \"redhat-operators-sfvz2\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.720574 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75r4x\" (UniqueName: \"kubernetes.io/projected/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-kube-api-access-75r4x\") pod \"redhat-operators-sfvz2\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:57 crc kubenswrapper[4871]: I1007 22:40:57.808830 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:40:58 crc kubenswrapper[4871]: I1007 22:40:58.374149 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sfvz2"] Oct 07 22:40:58 crc kubenswrapper[4871]: I1007 22:40:58.629472 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvz2" event={"ID":"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb","Type":"ContainerStarted","Data":"89b608ef0a5d12d1b784c62b92692a08702bed69f5965cef9f2bf659436aa0cb"} Oct 07 22:40:59 crc kubenswrapper[4871]: I1007 22:40:59.644786 4871 generic.go:334] "Generic (PLEG): container finished" podID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerID="c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe" exitCode=0 Oct 07 22:40:59 crc kubenswrapper[4871]: I1007 22:40:59.644918 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvz2" event={"ID":"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb","Type":"ContainerDied","Data":"c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe"} Oct 07 22:40:59 crc kubenswrapper[4871]: I1007 22:40:59.648619 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:41:00 crc kubenswrapper[4871]: I1007 22:41:00.654687 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvz2" event={"ID":"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb","Type":"ContainerStarted","Data":"0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3"} Oct 07 22:41:01 crc kubenswrapper[4871]: I1007 22:41:01.665045 4871 generic.go:334] "Generic (PLEG): container finished" podID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerID="0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3" exitCode=0 Oct 07 22:41:01 crc kubenswrapper[4871]: I1007 22:41:01.665129 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvz2" event={"ID":"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb","Type":"ContainerDied","Data":"0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3"} Oct 07 22:41:02 crc kubenswrapper[4871]: I1007 22:41:02.679160 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvz2" event={"ID":"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb","Type":"ContainerStarted","Data":"88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9"} Oct 07 22:41:02 crc kubenswrapper[4871]: I1007 22:41:02.708891 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sfvz2" podStartSLOduration=3.216003221 podStartE2EDuration="5.70886371s" podCreationTimestamp="2025-10-07 22:40:57 +0000 UTC" firstStartedPulling="2025-10-07 22:40:59.648309643 +0000 UTC m=+1933.451007716" lastFinishedPulling="2025-10-07 22:41:02.141170132 +0000 UTC m=+1935.943868205" observedRunningTime="2025-10-07 22:41:02.70165358 +0000 UTC m=+1936.504351663" watchObservedRunningTime="2025-10-07 22:41:02.70886371 +0000 UTC m=+1936.511561793" Oct 07 22:41:07 crc kubenswrapper[4871]: I1007 22:41:07.809189 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:41:07 crc kubenswrapper[4871]: I1007 22:41:07.809729 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:41:08 crc kubenswrapper[4871]: I1007 22:41:08.912178 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sfvz2" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerName="registry-server" probeResult="failure" output=< Oct 07 22:41:08 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 22:41:08 crc kubenswrapper[4871]: > Oct 07 22:41:17 crc kubenswrapper[4871]: I1007 22:41:17.887456 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:41:17 crc kubenswrapper[4871]: I1007 22:41:17.940434 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:41:18 crc kubenswrapper[4871]: I1007 22:41:18.126446 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sfvz2"] Oct 07 22:41:19 crc kubenswrapper[4871]: I1007 22:41:19.834585 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sfvz2" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerName="registry-server" containerID="cri-o://88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9" gracePeriod=2 Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.277432 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.465955 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75r4x\" (UniqueName: \"kubernetes.io/projected/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-kube-api-access-75r4x\") pod \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.466098 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-utilities\") pod \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.466226 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-catalog-content\") pod \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\" (UID: \"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb\") " Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.467020 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-utilities" (OuterVolumeSpecName: "utilities") pod "4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" (UID: "4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.472887 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-kube-api-access-75r4x" (OuterVolumeSpecName: "kube-api-access-75r4x") pod "4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" (UID: "4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb"). InnerVolumeSpecName "kube-api-access-75r4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.566524 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" (UID: "4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.567780 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75r4x\" (UniqueName: \"kubernetes.io/projected/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-kube-api-access-75r4x\") on node \"crc\" DevicePath \"\"" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.567832 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.567845 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.846137 4871 generic.go:334] "Generic (PLEG): container finished" podID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerID="88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9" exitCode=0 Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.846209 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvz2" event={"ID":"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb","Type":"ContainerDied","Data":"88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9"} Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.846259 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfvz2" event={"ID":"4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb","Type":"ContainerDied","Data":"89b608ef0a5d12d1b784c62b92692a08702bed69f5965cef9f2bf659436aa0cb"} Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.846286 4871 scope.go:117] "RemoveContainer" containerID="88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.846214 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfvz2" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.890535 4871 scope.go:117] "RemoveContainer" containerID="0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.899486 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sfvz2"] Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.906954 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sfvz2"] Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.931948 4871 scope.go:117] "RemoveContainer" containerID="c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.948979 4871 scope.go:117] "RemoveContainer" containerID="88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9" Oct 07 22:41:20 crc kubenswrapper[4871]: E1007 22:41:20.950093 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9\": container with ID starting with 88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9 not found: ID does not exist" containerID="88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.950203 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9"} err="failed to get container status \"88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9\": rpc error: code = NotFound desc = could not find container \"88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9\": container with ID starting with 88777ea849733671eec7d77d38f4a226d710ea10cabfa7b6309b9518fdfb2fc9 not found: ID does not exist" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.950251 4871 scope.go:117] "RemoveContainer" containerID="0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3" Oct 07 22:41:20 crc kubenswrapper[4871]: E1007 22:41:20.951302 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3\": container with ID starting with 0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3 not found: ID does not exist" containerID="0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.951439 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3"} err="failed to get container status \"0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3\": rpc error: code = NotFound desc = could not find container \"0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3\": container with ID starting with 0edcde813aa7804dcda91a043ef315d127ab73ae80faa8d4c9a554a14b04fce3 not found: ID does not exist" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.951576 4871 scope.go:117] "RemoveContainer" containerID="c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe" Oct 07 22:41:20 crc kubenswrapper[4871]: E1007 22:41:20.951994 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe\": container with ID starting with c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe not found: ID does not exist" containerID="c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.952035 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe"} err="failed to get container status \"c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe\": rpc error: code = NotFound desc = could not find container \"c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe\": container with ID starting with c83dd807acdd59b6833b49544649a28232ce609441fea9e3bbfd6f1ecc8646fe not found: ID does not exist" Oct 07 22:41:20 crc kubenswrapper[4871]: I1007 22:41:20.995956 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" path="/var/lib/kubelet/pods/4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb/volumes" Oct 07 22:41:34 crc kubenswrapper[4871]: I1007 22:41:34.846646 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kg44f"] Oct 07 22:41:34 crc kubenswrapper[4871]: E1007 22:41:34.847936 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerName="registry-server" Oct 07 22:41:34 crc kubenswrapper[4871]: I1007 22:41:34.847958 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerName="registry-server" Oct 07 22:41:34 crc kubenswrapper[4871]: E1007 22:41:34.847976 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerName="extract-utilities" Oct 07 22:41:34 crc kubenswrapper[4871]: I1007 22:41:34.847989 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerName="extract-utilities" Oct 07 22:41:34 crc kubenswrapper[4871]: E1007 22:41:34.848028 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerName="extract-content" Oct 07 22:41:34 crc kubenswrapper[4871]: I1007 22:41:34.848041 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerName="extract-content" Oct 07 22:41:34 crc kubenswrapper[4871]: I1007 22:41:34.848335 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c2d7d32-27c4-4d9d-9c8f-5b534d317cdb" containerName="registry-server" Oct 07 22:41:34 crc kubenswrapper[4871]: I1007 22:41:34.850455 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:34 crc kubenswrapper[4871]: I1007 22:41:34.901115 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg44f"] Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.001280 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h45fh\" (UniqueName: \"kubernetes.io/projected/e021e74a-b39a-40ae-a558-f61d48acd80c-kube-api-access-h45fh\") pod \"redhat-marketplace-kg44f\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.001489 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-catalog-content\") pod \"redhat-marketplace-kg44f\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.001705 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-utilities\") pod \"redhat-marketplace-kg44f\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.103739 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h45fh\" (UniqueName: \"kubernetes.io/projected/e021e74a-b39a-40ae-a558-f61d48acd80c-kube-api-access-h45fh\") pod \"redhat-marketplace-kg44f\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.103870 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-catalog-content\") pod \"redhat-marketplace-kg44f\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.103939 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-utilities\") pod \"redhat-marketplace-kg44f\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.104503 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-utilities\") pod \"redhat-marketplace-kg44f\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.105554 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-catalog-content\") pod \"redhat-marketplace-kg44f\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.137868 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h45fh\" (UniqueName: \"kubernetes.io/projected/e021e74a-b39a-40ae-a558-f61d48acd80c-kube-api-access-h45fh\") pod \"redhat-marketplace-kg44f\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.226836 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.512539 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.512887 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:41:35 crc kubenswrapper[4871]: I1007 22:41:35.703439 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg44f"] Oct 07 22:41:35 crc kubenswrapper[4871]: W1007 22:41:35.740376 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode021e74a_b39a_40ae_a558_f61d48acd80c.slice/crio-b7938adbfa34cbe88d3857e7139dc2ae5f8df271a8482c8a4a8ac5e78dc1920c WatchSource:0}: Error finding container b7938adbfa34cbe88d3857e7139dc2ae5f8df271a8482c8a4a8ac5e78dc1920c: Status 404 returned error can't find the container with id b7938adbfa34cbe88d3857e7139dc2ae5f8df271a8482c8a4a8ac5e78dc1920c Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.002132 4871 generic.go:334] "Generic (PLEG): container finished" podID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerID="311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c" exitCode=0 Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.002506 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg44f" event={"ID":"e021e74a-b39a-40ae-a558-f61d48acd80c","Type":"ContainerDied","Data":"311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c"} Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.002536 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg44f" event={"ID":"e021e74a-b39a-40ae-a558-f61d48acd80c","Type":"ContainerStarted","Data":"b7938adbfa34cbe88d3857e7139dc2ae5f8df271a8482c8a4a8ac5e78dc1920c"} Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.238723 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8z9jx"] Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.240435 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.253562 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8z9jx"] Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.422522 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86l8c\" (UniqueName: \"kubernetes.io/projected/4d593431-e21f-4bd8-9fef-c9f50ba47d47-kube-api-access-86l8c\") pod \"certified-operators-8z9jx\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.422641 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-catalog-content\") pod \"certified-operators-8z9jx\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.422683 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-utilities\") pod \"certified-operators-8z9jx\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.524395 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-catalog-content\") pod \"certified-operators-8z9jx\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.524472 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-utilities\") pod \"certified-operators-8z9jx\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.524573 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86l8c\" (UniqueName: \"kubernetes.io/projected/4d593431-e21f-4bd8-9fef-c9f50ba47d47-kube-api-access-86l8c\") pod \"certified-operators-8z9jx\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.524899 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-catalog-content\") pod \"certified-operators-8z9jx\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.525193 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-utilities\") pod \"certified-operators-8z9jx\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.551892 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86l8c\" (UniqueName: \"kubernetes.io/projected/4d593431-e21f-4bd8-9fef-c9f50ba47d47-kube-api-access-86l8c\") pod \"certified-operators-8z9jx\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.565142 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:36 crc kubenswrapper[4871]: I1007 22:41:36.853096 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8z9jx"] Oct 07 22:41:37 crc kubenswrapper[4871]: I1007 22:41:37.012945 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z9jx" event={"ID":"4d593431-e21f-4bd8-9fef-c9f50ba47d47","Type":"ContainerStarted","Data":"48384edc4f9c3605399d2cbc8f268957a24cbdf597626226778c0ce28509c8d1"} Oct 07 22:41:38 crc kubenswrapper[4871]: I1007 22:41:38.026107 4871 generic.go:334] "Generic (PLEG): container finished" podID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerID="01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c" exitCode=0 Oct 07 22:41:38 crc kubenswrapper[4871]: I1007 22:41:38.026678 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z9jx" event={"ID":"4d593431-e21f-4bd8-9fef-c9f50ba47d47","Type":"ContainerDied","Data":"01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c"} Oct 07 22:41:38 crc kubenswrapper[4871]: I1007 22:41:38.032379 4871 generic.go:334] "Generic (PLEG): container finished" podID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerID="38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d" exitCode=0 Oct 07 22:41:38 crc kubenswrapper[4871]: I1007 22:41:38.032476 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg44f" event={"ID":"e021e74a-b39a-40ae-a558-f61d48acd80c","Type":"ContainerDied","Data":"38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d"} Oct 07 22:41:40 crc kubenswrapper[4871]: I1007 22:41:40.058019 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg44f" event={"ID":"e021e74a-b39a-40ae-a558-f61d48acd80c","Type":"ContainerStarted","Data":"91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f"} Oct 07 22:41:41 crc kubenswrapper[4871]: I1007 22:41:41.068330 4871 generic.go:334] "Generic (PLEG): container finished" podID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerID="2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6" exitCode=0 Oct 07 22:41:41 crc kubenswrapper[4871]: I1007 22:41:41.068450 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z9jx" event={"ID":"4d593431-e21f-4bd8-9fef-c9f50ba47d47","Type":"ContainerDied","Data":"2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6"} Oct 07 22:41:41 crc kubenswrapper[4871]: I1007 22:41:41.104127 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kg44f" podStartSLOduration=3.323612438 podStartE2EDuration="7.104111498s" podCreationTimestamp="2025-10-07 22:41:34 +0000 UTC" firstStartedPulling="2025-10-07 22:41:36.005511394 +0000 UTC m=+1969.808209507" lastFinishedPulling="2025-10-07 22:41:39.786010454 +0000 UTC m=+1973.588708567" observedRunningTime="2025-10-07 22:41:41.102964217 +0000 UTC m=+1974.905662290" watchObservedRunningTime="2025-10-07 22:41:41.104111498 +0000 UTC m=+1974.906809571" Oct 07 22:41:42 crc kubenswrapper[4871]: I1007 22:41:42.081642 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z9jx" event={"ID":"4d593431-e21f-4bd8-9fef-c9f50ba47d47","Type":"ContainerStarted","Data":"f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443"} Oct 07 22:41:42 crc kubenswrapper[4871]: I1007 22:41:42.107437 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8z9jx" podStartSLOduration=2.6046218469999998 podStartE2EDuration="6.107416061s" podCreationTimestamp="2025-10-07 22:41:36 +0000 UTC" firstStartedPulling="2025-10-07 22:41:38.030519709 +0000 UTC m=+1971.833217822" lastFinishedPulling="2025-10-07 22:41:41.533313943 +0000 UTC m=+1975.336012036" observedRunningTime="2025-10-07 22:41:42.101268448 +0000 UTC m=+1975.903966551" watchObservedRunningTime="2025-10-07 22:41:42.107416061 +0000 UTC m=+1975.910114144" Oct 07 22:41:45 crc kubenswrapper[4871]: I1007 22:41:45.228304 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:45 crc kubenswrapper[4871]: I1007 22:41:45.228866 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:45 crc kubenswrapper[4871]: I1007 22:41:45.302485 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:46 crc kubenswrapper[4871]: I1007 22:41:46.175864 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:46 crc kubenswrapper[4871]: I1007 22:41:46.231822 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg44f"] Oct 07 22:41:46 crc kubenswrapper[4871]: I1007 22:41:46.565823 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:46 crc kubenswrapper[4871]: I1007 22:41:46.565896 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:46 crc kubenswrapper[4871]: I1007 22:41:46.639761 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:47 crc kubenswrapper[4871]: I1007 22:41:47.204345 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.138709 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kg44f" podUID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerName="registry-server" containerID="cri-o://91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f" gracePeriod=2 Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.230013 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8z9jx"] Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.641541 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.736634 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-utilities\") pod \"e021e74a-b39a-40ae-a558-f61d48acd80c\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.736884 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-catalog-content\") pod \"e021e74a-b39a-40ae-a558-f61d48acd80c\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.737085 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h45fh\" (UniqueName: \"kubernetes.io/projected/e021e74a-b39a-40ae-a558-f61d48acd80c-kube-api-access-h45fh\") pod \"e021e74a-b39a-40ae-a558-f61d48acd80c\" (UID: \"e021e74a-b39a-40ae-a558-f61d48acd80c\") " Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.738930 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-utilities" (OuterVolumeSpecName: "utilities") pod "e021e74a-b39a-40ae-a558-f61d48acd80c" (UID: "e021e74a-b39a-40ae-a558-f61d48acd80c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.743372 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e021e74a-b39a-40ae-a558-f61d48acd80c-kube-api-access-h45fh" (OuterVolumeSpecName: "kube-api-access-h45fh") pod "e021e74a-b39a-40ae-a558-f61d48acd80c" (UID: "e021e74a-b39a-40ae-a558-f61d48acd80c"). InnerVolumeSpecName "kube-api-access-h45fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.753962 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e021e74a-b39a-40ae-a558-f61d48acd80c" (UID: "e021e74a-b39a-40ae-a558-f61d48acd80c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.838860 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.838918 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h45fh\" (UniqueName: \"kubernetes.io/projected/e021e74a-b39a-40ae-a558-f61d48acd80c-kube-api-access-h45fh\") on node \"crc\" DevicePath \"\"" Oct 07 22:41:48 crc kubenswrapper[4871]: I1007 22:41:48.838940 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e021e74a-b39a-40ae-a558-f61d48acd80c-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.153351 4871 generic.go:334] "Generic (PLEG): container finished" podID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerID="91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f" exitCode=0 Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.153463 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg44f" event={"ID":"e021e74a-b39a-40ae-a558-f61d48acd80c","Type":"ContainerDied","Data":"91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f"} Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.153591 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kg44f" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.153886 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg44f" event={"ID":"e021e74a-b39a-40ae-a558-f61d48acd80c","Type":"ContainerDied","Data":"b7938adbfa34cbe88d3857e7139dc2ae5f8df271a8482c8a4a8ac5e78dc1920c"} Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.153965 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8z9jx" podUID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerName="registry-server" containerID="cri-o://f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443" gracePeriod=2 Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.154521 4871 scope.go:117] "RemoveContainer" containerID="91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.193832 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg44f"] Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.196013 4871 scope.go:117] "RemoveContainer" containerID="38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.206083 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg44f"] Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.225862 4871 scope.go:117] "RemoveContainer" containerID="311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.350920 4871 scope.go:117] "RemoveContainer" containerID="91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f" Oct 07 22:41:49 crc kubenswrapper[4871]: E1007 22:41:49.351621 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f\": container with ID starting with 91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f not found: ID does not exist" containerID="91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.351690 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f"} err="failed to get container status \"91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f\": rpc error: code = NotFound desc = could not find container \"91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f\": container with ID starting with 91b3c8e74f6bd8bc29b63d189c9a2605ae95c56484ae720d5e38a17ec314e96f not found: ID does not exist" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.351735 4871 scope.go:117] "RemoveContainer" containerID="38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d" Oct 07 22:41:49 crc kubenswrapper[4871]: E1007 22:41:49.352488 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d\": container with ID starting with 38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d not found: ID does not exist" containerID="38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.352548 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d"} err="failed to get container status \"38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d\": rpc error: code = NotFound desc = could not find container \"38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d\": container with ID starting with 38e65edf1767eb1f27f1e7274601fec2a17b6f5fac8485c7887cf784151a208d not found: ID does not exist" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.352595 4871 scope.go:117] "RemoveContainer" containerID="311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c" Oct 07 22:41:49 crc kubenswrapper[4871]: E1007 22:41:49.353185 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c\": container with ID starting with 311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c not found: ID does not exist" containerID="311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.353234 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c"} err="failed to get container status \"311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c\": rpc error: code = NotFound desc = could not find container \"311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c\": container with ID starting with 311734b918fca93e0ab18b910d3da6c3962647cb84458babef575150f33dbe6c not found: ID does not exist" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.704866 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.860224 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-utilities\") pod \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.860286 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-catalog-content\") pod \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.860334 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86l8c\" (UniqueName: \"kubernetes.io/projected/4d593431-e21f-4bd8-9fef-c9f50ba47d47-kube-api-access-86l8c\") pod \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\" (UID: \"4d593431-e21f-4bd8-9fef-c9f50ba47d47\") " Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.861621 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-utilities" (OuterVolumeSpecName: "utilities") pod "4d593431-e21f-4bd8-9fef-c9f50ba47d47" (UID: "4d593431-e21f-4bd8-9fef-c9f50ba47d47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.866944 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d593431-e21f-4bd8-9fef-c9f50ba47d47-kube-api-access-86l8c" (OuterVolumeSpecName: "kube-api-access-86l8c") pod "4d593431-e21f-4bd8-9fef-c9f50ba47d47" (UID: "4d593431-e21f-4bd8-9fef-c9f50ba47d47"). InnerVolumeSpecName "kube-api-access-86l8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.949360 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d593431-e21f-4bd8-9fef-c9f50ba47d47" (UID: "4d593431-e21f-4bd8-9fef-c9f50ba47d47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.962207 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.962264 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d593431-e21f-4bd8-9fef-c9f50ba47d47-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:41:49 crc kubenswrapper[4871]: I1007 22:41:49.962289 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86l8c\" (UniqueName: \"kubernetes.io/projected/4d593431-e21f-4bd8-9fef-c9f50ba47d47-kube-api-access-86l8c\") on node \"crc\" DevicePath \"\"" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.165255 4871 generic.go:334] "Generic (PLEG): container finished" podID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerID="f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443" exitCode=0 Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.165349 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z9jx" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.165334 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z9jx" event={"ID":"4d593431-e21f-4bd8-9fef-c9f50ba47d47","Type":"ContainerDied","Data":"f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443"} Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.165436 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z9jx" event={"ID":"4d593431-e21f-4bd8-9fef-c9f50ba47d47","Type":"ContainerDied","Data":"48384edc4f9c3605399d2cbc8f268957a24cbdf597626226778c0ce28509c8d1"} Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.165506 4871 scope.go:117] "RemoveContainer" containerID="f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.200576 4871 scope.go:117] "RemoveContainer" containerID="2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.225097 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8z9jx"] Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.235125 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8z9jx"] Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.246965 4871 scope.go:117] "RemoveContainer" containerID="01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.277689 4871 scope.go:117] "RemoveContainer" containerID="f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443" Oct 07 22:41:50 crc kubenswrapper[4871]: E1007 22:41:50.278243 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443\": container with ID starting with f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443 not found: ID does not exist" containerID="f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.278305 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443"} err="failed to get container status \"f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443\": rpc error: code = NotFound desc = could not find container \"f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443\": container with ID starting with f212da64b2542e9fc7d439e4d9beb7b23223a5c7dd89165209e2503fccbd3443 not found: ID does not exist" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.278341 4871 scope.go:117] "RemoveContainer" containerID="2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6" Oct 07 22:41:50 crc kubenswrapper[4871]: E1007 22:41:50.278919 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6\": container with ID starting with 2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6 not found: ID does not exist" containerID="2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.278963 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6"} err="failed to get container status \"2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6\": rpc error: code = NotFound desc = could not find container \"2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6\": container with ID starting with 2d983e4a1b800c565fa8479d60023e4b78b11ada75e63b8e56b840a44fc678e6 not found: ID does not exist" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.278991 4871 scope.go:117] "RemoveContainer" containerID="01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c" Oct 07 22:41:50 crc kubenswrapper[4871]: E1007 22:41:50.279621 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c\": container with ID starting with 01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c not found: ID does not exist" containerID="01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.279664 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c"} err="failed to get container status \"01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c\": rpc error: code = NotFound desc = could not find container \"01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c\": container with ID starting with 01ff104a62dc5e0ee313485a08087467d3a1f8faaef3fbce9f72f98d35188e4c not found: ID does not exist" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.998349 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" path="/var/lib/kubelet/pods/4d593431-e21f-4bd8-9fef-c9f50ba47d47/volumes" Oct 07 22:41:50 crc kubenswrapper[4871]: I1007 22:41:50.999782 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e021e74a-b39a-40ae-a558-f61d48acd80c" path="/var/lib/kubelet/pods/e021e74a-b39a-40ae-a558-f61d48acd80c/volumes" Oct 07 22:42:05 crc kubenswrapper[4871]: I1007 22:42:05.512281 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:42:05 crc kubenswrapper[4871]: I1007 22:42:05.512851 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.649764 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwx6t"] Oct 07 22:42:08 crc kubenswrapper[4871]: E1007 22:42:08.650676 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerName="extract-content" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.650762 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerName="extract-content" Oct 07 22:42:08 crc kubenswrapper[4871]: E1007 22:42:08.650783 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerName="registry-server" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.650824 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerName="registry-server" Oct 07 22:42:08 crc kubenswrapper[4871]: E1007 22:42:08.650847 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerName="extract-content" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.650860 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerName="extract-content" Oct 07 22:42:08 crc kubenswrapper[4871]: E1007 22:42:08.650893 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerName="extract-utilities" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.650905 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerName="extract-utilities" Oct 07 22:42:08 crc kubenswrapper[4871]: E1007 22:42:08.650931 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerName="registry-server" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.650943 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerName="registry-server" Oct 07 22:42:08 crc kubenswrapper[4871]: E1007 22:42:08.650980 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerName="extract-utilities" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.650992 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerName="extract-utilities" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.651309 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d593431-e21f-4bd8-9fef-c9f50ba47d47" containerName="registry-server" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.651346 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e021e74a-b39a-40ae-a558-f61d48acd80c" containerName="registry-server" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.653450 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.666634 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwx6t"] Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.718298 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-utilities\") pod \"community-operators-nwx6t\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.718393 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-catalog-content\") pod \"community-operators-nwx6t\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.718427 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfnnt\" (UniqueName: \"kubernetes.io/projected/42bf918b-fddf-4c54-8fb6-693cfe477984-kube-api-access-xfnnt\") pod \"community-operators-nwx6t\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.820888 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-utilities\") pod \"community-operators-nwx6t\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.820976 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-catalog-content\") pod \"community-operators-nwx6t\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.821002 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfnnt\" (UniqueName: \"kubernetes.io/projected/42bf918b-fddf-4c54-8fb6-693cfe477984-kube-api-access-xfnnt\") pod \"community-operators-nwx6t\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.822003 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-utilities\") pod \"community-operators-nwx6t\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.822028 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-catalog-content\") pod \"community-operators-nwx6t\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.843133 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfnnt\" (UniqueName: \"kubernetes.io/projected/42bf918b-fddf-4c54-8fb6-693cfe477984-kube-api-access-xfnnt\") pod \"community-operators-nwx6t\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:08 crc kubenswrapper[4871]: I1007 22:42:08.980323 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:09 crc kubenswrapper[4871]: I1007 22:42:09.482568 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwx6t"] Oct 07 22:42:10 crc kubenswrapper[4871]: I1007 22:42:10.341748 4871 generic.go:334] "Generic (PLEG): container finished" podID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerID="fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee" exitCode=0 Oct 07 22:42:10 crc kubenswrapper[4871]: I1007 22:42:10.341846 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwx6t" event={"ID":"42bf918b-fddf-4c54-8fb6-693cfe477984","Type":"ContainerDied","Data":"fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee"} Oct 07 22:42:10 crc kubenswrapper[4871]: I1007 22:42:10.342088 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwx6t" event={"ID":"42bf918b-fddf-4c54-8fb6-693cfe477984","Type":"ContainerStarted","Data":"ea49c33b8131f9a5cab9856af57fb97f2b6fcf7696199cb469fbc6bf322439b9"} Oct 07 22:42:11 crc kubenswrapper[4871]: I1007 22:42:11.355386 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwx6t" event={"ID":"42bf918b-fddf-4c54-8fb6-693cfe477984","Type":"ContainerStarted","Data":"14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de"} Oct 07 22:42:12 crc kubenswrapper[4871]: I1007 22:42:12.367459 4871 generic.go:334] "Generic (PLEG): container finished" podID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerID="14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de" exitCode=0 Oct 07 22:42:12 crc kubenswrapper[4871]: I1007 22:42:12.367516 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwx6t" event={"ID":"42bf918b-fddf-4c54-8fb6-693cfe477984","Type":"ContainerDied","Data":"14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de"} Oct 07 22:42:13 crc kubenswrapper[4871]: I1007 22:42:13.376746 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwx6t" event={"ID":"42bf918b-fddf-4c54-8fb6-693cfe477984","Type":"ContainerStarted","Data":"79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0"} Oct 07 22:42:13 crc kubenswrapper[4871]: I1007 22:42:13.404524 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwx6t" podStartSLOduration=2.949274678 podStartE2EDuration="5.404498134s" podCreationTimestamp="2025-10-07 22:42:08 +0000 UTC" firstStartedPulling="2025-10-07 22:42:10.346118739 +0000 UTC m=+2004.148816822" lastFinishedPulling="2025-10-07 22:42:12.801342195 +0000 UTC m=+2006.604040278" observedRunningTime="2025-10-07 22:42:13.401917615 +0000 UTC m=+2007.204615688" watchObservedRunningTime="2025-10-07 22:42:13.404498134 +0000 UTC m=+2007.207196247" Oct 07 22:42:18 crc kubenswrapper[4871]: I1007 22:42:18.980493 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:18 crc kubenswrapper[4871]: I1007 22:42:18.981584 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:19 crc kubenswrapper[4871]: I1007 22:42:19.029936 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:19 crc kubenswrapper[4871]: I1007 22:42:19.518585 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:23 crc kubenswrapper[4871]: I1007 22:42:23.830859 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwx6t"] Oct 07 22:42:23 crc kubenswrapper[4871]: I1007 22:42:23.832134 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwx6t" podUID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerName="registry-server" containerID="cri-o://79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0" gracePeriod=2 Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.361579 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.493855 4871 generic.go:334] "Generic (PLEG): container finished" podID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerID="79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0" exitCode=0 Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.493898 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwx6t" event={"ID":"42bf918b-fddf-4c54-8fb6-693cfe477984","Type":"ContainerDied","Data":"79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0"} Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.493922 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwx6t" event={"ID":"42bf918b-fddf-4c54-8fb6-693cfe477984","Type":"ContainerDied","Data":"ea49c33b8131f9a5cab9856af57fb97f2b6fcf7696199cb469fbc6bf322439b9"} Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.493939 4871 scope.go:117] "RemoveContainer" containerID="79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.493939 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwx6t" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.518355 4871 scope.go:117] "RemoveContainer" containerID="14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.542061 4871 scope.go:117] "RemoveContainer" containerID="fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.561789 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfnnt\" (UniqueName: \"kubernetes.io/projected/42bf918b-fddf-4c54-8fb6-693cfe477984-kube-api-access-xfnnt\") pod \"42bf918b-fddf-4c54-8fb6-693cfe477984\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.562179 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-catalog-content\") pod \"42bf918b-fddf-4c54-8fb6-693cfe477984\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.562255 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-utilities\") pod \"42bf918b-fddf-4c54-8fb6-693cfe477984\" (UID: \"42bf918b-fddf-4c54-8fb6-693cfe477984\") " Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.563402 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-utilities" (OuterVolumeSpecName: "utilities") pod "42bf918b-fddf-4c54-8fb6-693cfe477984" (UID: "42bf918b-fddf-4c54-8fb6-693cfe477984"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.568476 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42bf918b-fddf-4c54-8fb6-693cfe477984-kube-api-access-xfnnt" (OuterVolumeSpecName: "kube-api-access-xfnnt") pod "42bf918b-fddf-4c54-8fb6-693cfe477984" (UID: "42bf918b-fddf-4c54-8fb6-693cfe477984"). InnerVolumeSpecName "kube-api-access-xfnnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.581229 4871 scope.go:117] "RemoveContainer" containerID="79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0" Oct 07 22:42:24 crc kubenswrapper[4871]: E1007 22:42:24.582840 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0\": container with ID starting with 79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0 not found: ID does not exist" containerID="79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.582917 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0"} err="failed to get container status \"79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0\": rpc error: code = NotFound desc = could not find container \"79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0\": container with ID starting with 79d338610da87d0f972e7ba50adc23fd3c9f8177750468a865c6131122adedd0 not found: ID does not exist" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.582969 4871 scope.go:117] "RemoveContainer" containerID="14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de" Oct 07 22:42:24 crc kubenswrapper[4871]: E1007 22:42:24.583421 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de\": container with ID starting with 14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de not found: ID does not exist" containerID="14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.583503 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de"} err="failed to get container status \"14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de\": rpc error: code = NotFound desc = could not find container \"14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de\": container with ID starting with 14fe24143cb4f9daaa85379eb7b91ff70955a5e3fe1549ce7976aeeb8d6b21de not found: ID does not exist" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.583547 4871 scope.go:117] "RemoveContainer" containerID="fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee" Oct 07 22:42:24 crc kubenswrapper[4871]: E1007 22:42:24.584156 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee\": container with ID starting with fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee not found: ID does not exist" containerID="fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.584203 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee"} err="failed to get container status \"fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee\": rpc error: code = NotFound desc = could not find container \"fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee\": container with ID starting with fbcbfee87e07941d10e266011953e1488c97feec475f989aec724453c842dcee not found: ID does not exist" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.634276 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42bf918b-fddf-4c54-8fb6-693cfe477984" (UID: "42bf918b-fddf-4c54-8fb6-693cfe477984"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.664004 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.664084 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42bf918b-fddf-4c54-8fb6-693cfe477984-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.664106 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfnnt\" (UniqueName: \"kubernetes.io/projected/42bf918b-fddf-4c54-8fb6-693cfe477984-kube-api-access-xfnnt\") on node \"crc\" DevicePath \"\"" Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.857265 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwx6t"] Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.871748 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwx6t"] Oct 07 22:42:24 crc kubenswrapper[4871]: I1007 22:42:24.993263 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42bf918b-fddf-4c54-8fb6-693cfe477984" path="/var/lib/kubelet/pods/42bf918b-fddf-4c54-8fb6-693cfe477984/volumes" Oct 07 22:42:35 crc kubenswrapper[4871]: I1007 22:42:35.512055 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:42:35 crc kubenswrapper[4871]: I1007 22:42:35.512938 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:42:35 crc kubenswrapper[4871]: I1007 22:42:35.513012 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:42:35 crc kubenswrapper[4871]: I1007 22:42:35.513983 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ecf3f7625fa1b5db1c3a7b56d23927b826eefe23c4c9f7a906222c77a0c97263"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:42:35 crc kubenswrapper[4871]: I1007 22:42:35.514088 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://ecf3f7625fa1b5db1c3a7b56d23927b826eefe23c4c9f7a906222c77a0c97263" gracePeriod=600 Oct 07 22:42:36 crc kubenswrapper[4871]: I1007 22:42:36.608684 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="ecf3f7625fa1b5db1c3a7b56d23927b826eefe23c4c9f7a906222c77a0c97263" exitCode=0 Oct 07 22:42:36 crc kubenswrapper[4871]: I1007 22:42:36.608847 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"ecf3f7625fa1b5db1c3a7b56d23927b826eefe23c4c9f7a906222c77a0c97263"} Oct 07 22:42:36 crc kubenswrapper[4871]: I1007 22:42:36.609762 4871 scope.go:117] "RemoveContainer" containerID="2de2269ba82ab1545b30ebdb613a7a56dab89d1d6faf2c518b40a9a9c5a9abf6" Oct 07 22:42:37 crc kubenswrapper[4871]: I1007 22:42:37.620924 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929"} Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.159386 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94"] Oct 07 22:45:00 crc kubenswrapper[4871]: E1007 22:45:00.160507 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerName="extract-utilities" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.160522 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerName="extract-utilities" Oct 07 22:45:00 crc kubenswrapper[4871]: E1007 22:45:00.160547 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerName="registry-server" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.160556 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerName="registry-server" Oct 07 22:45:00 crc kubenswrapper[4871]: E1007 22:45:00.160574 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerName="extract-content" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.160584 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerName="extract-content" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.160773 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="42bf918b-fddf-4c54-8fb6-693cfe477984" containerName="registry-server" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.161412 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.166167 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.166745 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.188566 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94"] Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.244598 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm8zt\" (UniqueName: \"kubernetes.io/projected/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-kube-api-access-rm8zt\") pod \"collect-profiles-29331285-8ss94\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.244698 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-secret-volume\") pod \"collect-profiles-29331285-8ss94\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.244772 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-config-volume\") pod \"collect-profiles-29331285-8ss94\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.346949 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm8zt\" (UniqueName: \"kubernetes.io/projected/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-kube-api-access-rm8zt\") pod \"collect-profiles-29331285-8ss94\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.347042 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-secret-volume\") pod \"collect-profiles-29331285-8ss94\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.347088 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-config-volume\") pod \"collect-profiles-29331285-8ss94\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.348182 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-config-volume\") pod \"collect-profiles-29331285-8ss94\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.361192 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-secret-volume\") pod \"collect-profiles-29331285-8ss94\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.372078 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm8zt\" (UniqueName: \"kubernetes.io/projected/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-kube-api-access-rm8zt\") pod \"collect-profiles-29331285-8ss94\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.498469 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:00 crc kubenswrapper[4871]: I1007 22:45:00.751882 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94"] Oct 07 22:45:01 crc kubenswrapper[4871]: I1007 22:45:01.003806 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" event={"ID":"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8","Type":"ContainerStarted","Data":"0c96262bf0a814a6f26dce4a1cf750095780f028692d0eedfcfefb7ffcaf3d79"} Oct 07 22:45:01 crc kubenswrapper[4871]: I1007 22:45:01.004514 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" event={"ID":"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8","Type":"ContainerStarted","Data":"b3ec721af4f736e2ddbe23c880a00e3c18a8dc38f41be961fbe21ca8e7cc6bd0"} Oct 07 22:45:01 crc kubenswrapper[4871]: I1007 22:45:01.026046 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" podStartSLOduration=1.026020686 podStartE2EDuration="1.026020686s" podCreationTimestamp="2025-10-07 22:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 22:45:01.023723985 +0000 UTC m=+2174.826422088" watchObservedRunningTime="2025-10-07 22:45:01.026020686 +0000 UTC m=+2174.828718759" Oct 07 22:45:02 crc kubenswrapper[4871]: I1007 22:45:02.017827 4871 generic.go:334] "Generic (PLEG): container finished" podID="eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8" containerID="0c96262bf0a814a6f26dce4a1cf750095780f028692d0eedfcfefb7ffcaf3d79" exitCode=0 Oct 07 22:45:02 crc kubenswrapper[4871]: I1007 22:45:02.017946 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" event={"ID":"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8","Type":"ContainerDied","Data":"0c96262bf0a814a6f26dce4a1cf750095780f028692d0eedfcfefb7ffcaf3d79"} Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.312262 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.490466 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-secret-volume\") pod \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.490617 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-config-volume\") pod \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.490731 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm8zt\" (UniqueName: \"kubernetes.io/projected/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-kube-api-access-rm8zt\") pod \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\" (UID: \"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8\") " Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.492005 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-config-volume" (OuterVolumeSpecName: "config-volume") pod "eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8" (UID: "eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.499963 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8" (UID: "eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.501311 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-kube-api-access-rm8zt" (OuterVolumeSpecName: "kube-api-access-rm8zt") pod "eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8" (UID: "eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8"). InnerVolumeSpecName "kube-api-access-rm8zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.602458 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.602516 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 22:45:03 crc kubenswrapper[4871]: I1007 22:45:03.602533 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm8zt\" (UniqueName: \"kubernetes.io/projected/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8-kube-api-access-rm8zt\") on node \"crc\" DevicePath \"\"" Oct 07 22:45:04 crc kubenswrapper[4871]: I1007 22:45:04.036366 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" event={"ID":"eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8","Type":"ContainerDied","Data":"b3ec721af4f736e2ddbe23c880a00e3c18a8dc38f41be961fbe21ca8e7cc6bd0"} Oct 07 22:45:04 crc kubenswrapper[4871]: I1007 22:45:04.036410 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3ec721af4f736e2ddbe23c880a00e3c18a8dc38f41be961fbe21ca8e7cc6bd0" Oct 07 22:45:04 crc kubenswrapper[4871]: I1007 22:45:04.036453 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94" Oct 07 22:45:04 crc kubenswrapper[4871]: I1007 22:45:04.395658 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw"] Oct 07 22:45:04 crc kubenswrapper[4871]: I1007 22:45:04.401295 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331240-6x9cw"] Oct 07 22:45:05 crc kubenswrapper[4871]: I1007 22:45:05.000554 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a916502-6bc9-48ba-906e-831e6fe85e3b" path="/var/lib/kubelet/pods/1a916502-6bc9-48ba-906e-831e6fe85e3b/volumes" Oct 07 22:45:05 crc kubenswrapper[4871]: I1007 22:45:05.512864 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:45:05 crc kubenswrapper[4871]: I1007 22:45:05.512973 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:45:35 crc kubenswrapper[4871]: I1007 22:45:35.512432 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:45:35 crc kubenswrapper[4871]: I1007 22:45:35.513265 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:45:54 crc kubenswrapper[4871]: I1007 22:45:54.422524 4871 scope.go:117] "RemoveContainer" containerID="7a551e5fadb1293eab5adb4685604c781f15988d0b22ba8462832dc5b987e10d" Oct 07 22:46:05 crc kubenswrapper[4871]: I1007 22:46:05.512706 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:46:05 crc kubenswrapper[4871]: I1007 22:46:05.513591 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:46:05 crc kubenswrapper[4871]: I1007 22:46:05.513672 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:46:05 crc kubenswrapper[4871]: I1007 22:46:05.514506 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:46:05 crc kubenswrapper[4871]: I1007 22:46:05.514602 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" gracePeriod=600 Oct 07 22:46:06 crc kubenswrapper[4871]: E1007 22:46:06.218527 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:46:06 crc kubenswrapper[4871]: I1007 22:46:06.595170 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" exitCode=0 Oct 07 22:46:06 crc kubenswrapper[4871]: I1007 22:46:06.595231 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929"} Oct 07 22:46:06 crc kubenswrapper[4871]: I1007 22:46:06.595284 4871 scope.go:117] "RemoveContainer" containerID="ecf3f7625fa1b5db1c3a7b56d23927b826eefe23c4c9f7a906222c77a0c97263" Oct 07 22:46:06 crc kubenswrapper[4871]: I1007 22:46:06.596070 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:46:06 crc kubenswrapper[4871]: E1007 22:46:06.596438 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:46:20 crc kubenswrapper[4871]: I1007 22:46:20.982782 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:46:20 crc kubenswrapper[4871]: E1007 22:46:20.983719 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:46:35 crc kubenswrapper[4871]: I1007 22:46:35.982624 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:46:35 crc kubenswrapper[4871]: E1007 22:46:35.983548 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:46:46 crc kubenswrapper[4871]: I1007 22:46:46.986062 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:46:46 crc kubenswrapper[4871]: E1007 22:46:46.986940 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:46:57 crc kubenswrapper[4871]: I1007 22:46:57.982538 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:46:57 crc kubenswrapper[4871]: E1007 22:46:57.983620 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:47:09 crc kubenswrapper[4871]: I1007 22:47:09.982770 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:47:09 crc kubenswrapper[4871]: E1007 22:47:09.985542 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:47:22 crc kubenswrapper[4871]: I1007 22:47:22.982889 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:47:22 crc kubenswrapper[4871]: E1007 22:47:22.983978 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:47:34 crc kubenswrapper[4871]: I1007 22:47:34.983516 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:47:34 crc kubenswrapper[4871]: E1007 22:47:34.984426 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:47:47 crc kubenswrapper[4871]: I1007 22:47:47.982902 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:47:47 crc kubenswrapper[4871]: E1007 22:47:47.983870 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:47:59 crc kubenswrapper[4871]: I1007 22:47:59.983024 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:47:59 crc kubenswrapper[4871]: E1007 22:47:59.984021 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:48:14 crc kubenswrapper[4871]: I1007 22:48:14.983954 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:48:14 crc kubenswrapper[4871]: E1007 22:48:14.985377 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:48:29 crc kubenswrapper[4871]: I1007 22:48:29.982179 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:48:29 crc kubenswrapper[4871]: E1007 22:48:29.983055 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:48:44 crc kubenswrapper[4871]: I1007 22:48:44.982878 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:48:44 crc kubenswrapper[4871]: E1007 22:48:44.983863 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:48:57 crc kubenswrapper[4871]: I1007 22:48:57.982621 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:48:57 crc kubenswrapper[4871]: E1007 22:48:57.983599 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:49:09 crc kubenswrapper[4871]: I1007 22:49:09.982891 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:49:09 crc kubenswrapper[4871]: E1007 22:49:09.984123 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:49:22 crc kubenswrapper[4871]: I1007 22:49:22.982307 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:49:22 crc kubenswrapper[4871]: E1007 22:49:22.983080 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:49:33 crc kubenswrapper[4871]: I1007 22:49:33.982991 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:49:33 crc kubenswrapper[4871]: E1007 22:49:33.984047 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:49:45 crc kubenswrapper[4871]: I1007 22:49:45.983091 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:49:45 crc kubenswrapper[4871]: E1007 22:49:45.984421 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:50:00 crc kubenswrapper[4871]: I1007 22:50:00.982683 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:50:00 crc kubenswrapper[4871]: E1007 22:50:00.983733 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:50:11 crc kubenswrapper[4871]: I1007 22:50:11.982235 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:50:11 crc kubenswrapper[4871]: E1007 22:50:11.983383 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:50:23 crc kubenswrapper[4871]: I1007 22:50:23.982411 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:50:23 crc kubenswrapper[4871]: E1007 22:50:23.983489 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:50:37 crc kubenswrapper[4871]: I1007 22:50:37.982689 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:50:37 crc kubenswrapper[4871]: E1007 22:50:37.984196 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:50:48 crc kubenswrapper[4871]: I1007 22:50:48.982720 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:50:48 crc kubenswrapper[4871]: E1007 22:50:48.983595 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:51:01 crc kubenswrapper[4871]: I1007 22:51:01.983150 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:51:01 crc kubenswrapper[4871]: E1007 22:51:01.984079 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:51:12 crc kubenswrapper[4871]: I1007 22:51:12.983168 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:51:13 crc kubenswrapper[4871]: I1007 22:51:13.600277 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"7ed4da1e7ab23d6ef0965429501276019d9fa05f06889389751d667e20f40e45"} Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.444310 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qmpkp"] Oct 07 22:52:18 crc kubenswrapper[4871]: E1007 22:52:18.446719 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8" containerName="collect-profiles" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.446747 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8" containerName="collect-profiles" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.448128 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8" containerName="collect-profiles" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.450987 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.464599 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qmpkp"] Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.636719 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-utilities\") pod \"certified-operators-qmpkp\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.637024 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqn4k\" (UniqueName: \"kubernetes.io/projected/69de2987-be53-4f93-8a3e-a91437e33bbd-kube-api-access-hqn4k\") pod \"certified-operators-qmpkp\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.637154 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-catalog-content\") pod \"certified-operators-qmpkp\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.738326 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-utilities\") pod \"certified-operators-qmpkp\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.738641 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqn4k\" (UniqueName: \"kubernetes.io/projected/69de2987-be53-4f93-8a3e-a91437e33bbd-kube-api-access-hqn4k\") pod \"certified-operators-qmpkp\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.738815 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-catalog-content\") pod \"certified-operators-qmpkp\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.738898 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-utilities\") pod \"certified-operators-qmpkp\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.739290 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-catalog-content\") pod \"certified-operators-qmpkp\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.771580 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqn4k\" (UniqueName: \"kubernetes.io/projected/69de2987-be53-4f93-8a3e-a91437e33bbd-kube-api-access-hqn4k\") pod \"certified-operators-qmpkp\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:18 crc kubenswrapper[4871]: I1007 22:52:18.781617 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:19 crc kubenswrapper[4871]: I1007 22:52:19.058257 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qmpkp"] Oct 07 22:52:19 crc kubenswrapper[4871]: I1007 22:52:19.245974 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmpkp" event={"ID":"69de2987-be53-4f93-8a3e-a91437e33bbd","Type":"ContainerStarted","Data":"93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec"} Oct 07 22:52:19 crc kubenswrapper[4871]: I1007 22:52:19.246021 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmpkp" event={"ID":"69de2987-be53-4f93-8a3e-a91437e33bbd","Type":"ContainerStarted","Data":"c34ab5b9e72b365f14b4d72d618c7cc82705c69e61aacd7d2d74517ee47825bf"} Oct 07 22:52:19 crc kubenswrapper[4871]: I1007 22:52:19.248827 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 22:52:20 crc kubenswrapper[4871]: I1007 22:52:20.256726 4871 generic.go:334] "Generic (PLEG): container finished" podID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerID="93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec" exitCode=0 Oct 07 22:52:20 crc kubenswrapper[4871]: I1007 22:52:20.256804 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmpkp" event={"ID":"69de2987-be53-4f93-8a3e-a91437e33bbd","Type":"ContainerDied","Data":"93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec"} Oct 07 22:52:20 crc kubenswrapper[4871]: I1007 22:52:20.259767 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmpkp" event={"ID":"69de2987-be53-4f93-8a3e-a91437e33bbd","Type":"ContainerStarted","Data":"f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e"} Oct 07 22:52:21 crc kubenswrapper[4871]: I1007 22:52:21.268997 4871 generic.go:334] "Generic (PLEG): container finished" podID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerID="f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e" exitCode=0 Oct 07 22:52:21 crc kubenswrapper[4871]: I1007 22:52:21.269332 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmpkp" event={"ID":"69de2987-be53-4f93-8a3e-a91437e33bbd","Type":"ContainerDied","Data":"f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e"} Oct 07 22:52:22 crc kubenswrapper[4871]: I1007 22:52:22.283176 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmpkp" event={"ID":"69de2987-be53-4f93-8a3e-a91437e33bbd","Type":"ContainerStarted","Data":"682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6"} Oct 07 22:52:22 crc kubenswrapper[4871]: I1007 22:52:22.307947 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qmpkp" podStartSLOduration=1.806153246 podStartE2EDuration="4.307925918s" podCreationTimestamp="2025-10-07 22:52:18 +0000 UTC" firstStartedPulling="2025-10-07 22:52:19.248572743 +0000 UTC m=+2613.051270816" lastFinishedPulling="2025-10-07 22:52:21.750345385 +0000 UTC m=+2615.553043488" observedRunningTime="2025-10-07 22:52:22.306708276 +0000 UTC m=+2616.109406399" watchObservedRunningTime="2025-10-07 22:52:22.307925918 +0000 UTC m=+2616.110624011" Oct 07 22:52:28 crc kubenswrapper[4871]: I1007 22:52:28.782745 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:28 crc kubenswrapper[4871]: I1007 22:52:28.783311 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:28 crc kubenswrapper[4871]: I1007 22:52:28.857677 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:29 crc kubenswrapper[4871]: I1007 22:52:29.442231 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:29 crc kubenswrapper[4871]: I1007 22:52:29.505910 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qmpkp"] Oct 07 22:52:31 crc kubenswrapper[4871]: I1007 22:52:31.388012 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qmpkp" podUID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerName="registry-server" containerID="cri-o://682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6" gracePeriod=2 Oct 07 22:52:31 crc kubenswrapper[4871]: I1007 22:52:31.943351 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.063229 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-utilities\") pod \"69de2987-be53-4f93-8a3e-a91437e33bbd\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.064420 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-catalog-content\") pod \"69de2987-be53-4f93-8a3e-a91437e33bbd\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.064490 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqn4k\" (UniqueName: \"kubernetes.io/projected/69de2987-be53-4f93-8a3e-a91437e33bbd-kube-api-access-hqn4k\") pod \"69de2987-be53-4f93-8a3e-a91437e33bbd\" (UID: \"69de2987-be53-4f93-8a3e-a91437e33bbd\") " Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.065268 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-utilities" (OuterVolumeSpecName: "utilities") pod "69de2987-be53-4f93-8a3e-a91437e33bbd" (UID: "69de2987-be53-4f93-8a3e-a91437e33bbd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.074227 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69de2987-be53-4f93-8a3e-a91437e33bbd-kube-api-access-hqn4k" (OuterVolumeSpecName: "kube-api-access-hqn4k") pod "69de2987-be53-4f93-8a3e-a91437e33bbd" (UID: "69de2987-be53-4f93-8a3e-a91437e33bbd"). InnerVolumeSpecName "kube-api-access-hqn4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.126523 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69de2987-be53-4f93-8a3e-a91437e33bbd" (UID: "69de2987-be53-4f93-8a3e-a91437e33bbd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.165710 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.165747 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqn4k\" (UniqueName: \"kubernetes.io/projected/69de2987-be53-4f93-8a3e-a91437e33bbd-kube-api-access-hqn4k\") on node \"crc\" DevicePath \"\"" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.166105 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69de2987-be53-4f93-8a3e-a91437e33bbd-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.401944 4871 generic.go:334] "Generic (PLEG): container finished" podID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerID="682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6" exitCode=0 Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.402009 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmpkp" event={"ID":"69de2987-be53-4f93-8a3e-a91437e33bbd","Type":"ContainerDied","Data":"682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6"} Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.402048 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmpkp" event={"ID":"69de2987-be53-4f93-8a3e-a91437e33bbd","Type":"ContainerDied","Data":"c34ab5b9e72b365f14b4d72d618c7cc82705c69e61aacd7d2d74517ee47825bf"} Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.402077 4871 scope.go:117] "RemoveContainer" containerID="682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.402245 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmpkp" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.434688 4871 scope.go:117] "RemoveContainer" containerID="f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.457234 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qmpkp"] Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.464450 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qmpkp"] Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.479512 4871 scope.go:117] "RemoveContainer" containerID="93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.508678 4871 scope.go:117] "RemoveContainer" containerID="682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6" Oct 07 22:52:32 crc kubenswrapper[4871]: E1007 22:52:32.510693 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6\": container with ID starting with 682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6 not found: ID does not exist" containerID="682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.510740 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6"} err="failed to get container status \"682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6\": rpc error: code = NotFound desc = could not find container \"682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6\": container with ID starting with 682e05979e92e3da5d0237e794399aeb8aa63533251b34eb75f68157fc3b6fb6 not found: ID does not exist" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.510767 4871 scope.go:117] "RemoveContainer" containerID="f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e" Oct 07 22:52:32 crc kubenswrapper[4871]: E1007 22:52:32.511308 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e\": container with ID starting with f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e not found: ID does not exist" containerID="f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.511341 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e"} err="failed to get container status \"f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e\": rpc error: code = NotFound desc = could not find container \"f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e\": container with ID starting with f1345b263e72b6870ddf89f4cd119ea2186e076dc1daac7c84ba68b4ea44f36e not found: ID does not exist" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.511361 4871 scope.go:117] "RemoveContainer" containerID="93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec" Oct 07 22:52:32 crc kubenswrapper[4871]: E1007 22:52:32.511655 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec\": container with ID starting with 93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec not found: ID does not exist" containerID="93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.511707 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec"} err="failed to get container status \"93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec\": rpc error: code = NotFound desc = could not find container \"93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec\": container with ID starting with 93c631e23c22c29c5e0f688a519f3b5f7c1fea15f6f8829c22945dcde4b56bec not found: ID does not exist" Oct 07 22:52:32 crc kubenswrapper[4871]: I1007 22:52:32.999129 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69de2987-be53-4f93-8a3e-a91437e33bbd" path="/var/lib/kubelet/pods/69de2987-be53-4f93-8a3e-a91437e33bbd/volumes" Oct 07 22:53:35 crc kubenswrapper[4871]: I1007 22:53:35.512561 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:53:35 crc kubenswrapper[4871]: I1007 22:53:35.513171 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:54:05 crc kubenswrapper[4871]: I1007 22:54:05.511972 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:54:05 crc kubenswrapper[4871]: I1007 22:54:05.512696 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:54:35 crc kubenswrapper[4871]: I1007 22:54:35.512130 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:54:35 crc kubenswrapper[4871]: I1007 22:54:35.512742 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:54:35 crc kubenswrapper[4871]: I1007 22:54:35.512838 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:54:35 crc kubenswrapper[4871]: I1007 22:54:35.513661 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ed4da1e7ab23d6ef0965429501276019d9fa05f06889389751d667e20f40e45"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:54:35 crc kubenswrapper[4871]: I1007 22:54:35.513832 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://7ed4da1e7ab23d6ef0965429501276019d9fa05f06889389751d667e20f40e45" gracePeriod=600 Oct 07 22:54:36 crc kubenswrapper[4871]: I1007 22:54:36.592196 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="7ed4da1e7ab23d6ef0965429501276019d9fa05f06889389751d667e20f40e45" exitCode=0 Oct 07 22:54:36 crc kubenswrapper[4871]: I1007 22:54:36.592280 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"7ed4da1e7ab23d6ef0965429501276019d9fa05f06889389751d667e20f40e45"} Oct 07 22:54:36 crc kubenswrapper[4871]: I1007 22:54:36.592870 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb"} Oct 07 22:54:36 crc kubenswrapper[4871]: I1007 22:54:36.592901 4871 scope.go:117] "RemoveContainer" containerID="647df072ace7f3fd6d832358283d9538fa2f069479729de4e485a96cff078929" Oct 07 22:56:35 crc kubenswrapper[4871]: I1007 22:56:35.511993 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:56:35 crc kubenswrapper[4871]: I1007 22:56:35.512776 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.317314 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l2tz6"] Oct 07 22:56:42 crc kubenswrapper[4871]: E1007 22:56:42.319428 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerName="extract-content" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.319539 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerName="extract-content" Oct 07 22:56:42 crc kubenswrapper[4871]: E1007 22:56:42.319646 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerName="extract-utilities" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.319727 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerName="extract-utilities" Oct 07 22:56:42 crc kubenswrapper[4871]: E1007 22:56:42.319859 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerName="registry-server" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.319951 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerName="registry-server" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.320198 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="69de2987-be53-4f93-8a3e-a91437e33bbd" containerName="registry-server" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.321492 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.347989 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l2tz6"] Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.493353 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-utilities\") pod \"redhat-operators-l2tz6\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.493743 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-catalog-content\") pod \"redhat-operators-l2tz6\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.493810 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xqjg\" (UniqueName: \"kubernetes.io/projected/ba306cf0-f483-4d8d-9712-863ef4db5773-kube-api-access-6xqjg\") pod \"redhat-operators-l2tz6\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.595203 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-utilities\") pod \"redhat-operators-l2tz6\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.595267 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-catalog-content\") pod \"redhat-operators-l2tz6\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.595288 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xqjg\" (UniqueName: \"kubernetes.io/projected/ba306cf0-f483-4d8d-9712-863ef4db5773-kube-api-access-6xqjg\") pod \"redhat-operators-l2tz6\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.595829 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-utilities\") pod \"redhat-operators-l2tz6\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.595851 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-catalog-content\") pod \"redhat-operators-l2tz6\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.614647 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xqjg\" (UniqueName: \"kubernetes.io/projected/ba306cf0-f483-4d8d-9712-863ef4db5773-kube-api-access-6xqjg\") pod \"redhat-operators-l2tz6\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.657991 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:42 crc kubenswrapper[4871]: I1007 22:56:42.930402 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l2tz6"] Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.311189 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l9szb"] Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.312538 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.321835 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9szb"] Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.406827 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-catalog-content\") pod \"community-operators-l9szb\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.407007 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2vzx\" (UniqueName: \"kubernetes.io/projected/847482fb-1339-4a87-9d4d-f178aaf0435f-kube-api-access-z2vzx\") pod \"community-operators-l9szb\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.407206 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-utilities\") pod \"community-operators-l9szb\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.508149 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-catalog-content\") pod \"community-operators-l9szb\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.508223 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2vzx\" (UniqueName: \"kubernetes.io/projected/847482fb-1339-4a87-9d4d-f178aaf0435f-kube-api-access-z2vzx\") pod \"community-operators-l9szb\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.508279 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-utilities\") pod \"community-operators-l9szb\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.508610 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-catalog-content\") pod \"community-operators-l9szb\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.508670 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-utilities\") pod \"community-operators-l9szb\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.531747 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2vzx\" (UniqueName: \"kubernetes.io/projected/847482fb-1339-4a87-9d4d-f178aaf0435f-kube-api-access-z2vzx\") pod \"community-operators-l9szb\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.637150 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.807001 4871 generic.go:334] "Generic (PLEG): container finished" podID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerID="61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc" exitCode=0 Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.807045 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2tz6" event={"ID":"ba306cf0-f483-4d8d-9712-863ef4db5773","Type":"ContainerDied","Data":"61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc"} Oct 07 22:56:43 crc kubenswrapper[4871]: I1007 22:56:43.807068 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2tz6" event={"ID":"ba306cf0-f483-4d8d-9712-863ef4db5773","Type":"ContainerStarted","Data":"dc34d63e32aa3dad8c5a034e2d12853d10b09d0883101e6646f401821b60c89d"} Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.114400 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9szb"] Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.725401 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lggxl"] Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.728371 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.733749 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f7a029e-6cb3-4051-b2b2-a393044a2ea3-catalog-content\") pod \"redhat-marketplace-lggxl\" (UID: \"8f7a029e-6cb3-4051-b2b2-a393044a2ea3\") " pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.733932 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6wft\" (UniqueName: \"kubernetes.io/projected/8f7a029e-6cb3-4051-b2b2-a393044a2ea3-kube-api-access-w6wft\") pod \"redhat-marketplace-lggxl\" (UID: \"8f7a029e-6cb3-4051-b2b2-a393044a2ea3\") " pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.734035 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f7a029e-6cb3-4051-b2b2-a393044a2ea3-utilities\") pod \"redhat-marketplace-lggxl\" (UID: \"8f7a029e-6cb3-4051-b2b2-a393044a2ea3\") " pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.739271 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lggxl"] Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.819572 4871 generic.go:334] "Generic (PLEG): container finished" podID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerID="56793ceef6a6908856d3879a985a2163e075bb1e0dd6d3328e8119ff4172e4e4" exitCode=0 Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.819624 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9szb" event={"ID":"847482fb-1339-4a87-9d4d-f178aaf0435f","Type":"ContainerDied","Data":"56793ceef6a6908856d3879a985a2163e075bb1e0dd6d3328e8119ff4172e4e4"} Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.819654 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9szb" event={"ID":"847482fb-1339-4a87-9d4d-f178aaf0435f","Type":"ContainerStarted","Data":"04d2c733a2c2b3ce19e3d9d4411873a9a9567d6e1c8e35272808966b23f61e25"} Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.835458 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6wft\" (UniqueName: \"kubernetes.io/projected/8f7a029e-6cb3-4051-b2b2-a393044a2ea3-kube-api-access-w6wft\") pod \"redhat-marketplace-lggxl\" (UID: \"8f7a029e-6cb3-4051-b2b2-a393044a2ea3\") " pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.835589 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f7a029e-6cb3-4051-b2b2-a393044a2ea3-utilities\") pod \"redhat-marketplace-lggxl\" (UID: \"8f7a029e-6cb3-4051-b2b2-a393044a2ea3\") " pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.835658 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f7a029e-6cb3-4051-b2b2-a393044a2ea3-catalog-content\") pod \"redhat-marketplace-lggxl\" (UID: \"8f7a029e-6cb3-4051-b2b2-a393044a2ea3\") " pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.837870 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f7a029e-6cb3-4051-b2b2-a393044a2ea3-catalog-content\") pod \"redhat-marketplace-lggxl\" (UID: \"8f7a029e-6cb3-4051-b2b2-a393044a2ea3\") " pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.838274 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f7a029e-6cb3-4051-b2b2-a393044a2ea3-utilities\") pod \"redhat-marketplace-lggxl\" (UID: \"8f7a029e-6cb3-4051-b2b2-a393044a2ea3\") " pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:44 crc kubenswrapper[4871]: I1007 22:56:44.870172 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6wft\" (UniqueName: \"kubernetes.io/projected/8f7a029e-6cb3-4051-b2b2-a393044a2ea3-kube-api-access-w6wft\") pod \"redhat-marketplace-lggxl\" (UID: \"8f7a029e-6cb3-4051-b2b2-a393044a2ea3\") " pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:45 crc kubenswrapper[4871]: I1007 22:56:45.061730 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:45 crc kubenswrapper[4871]: I1007 22:56:45.314038 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lggxl"] Oct 07 22:56:45 crc kubenswrapper[4871]: I1007 22:56:45.832458 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2tz6" event={"ID":"ba306cf0-f483-4d8d-9712-863ef4db5773","Type":"ContainerStarted","Data":"f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484"} Oct 07 22:56:45 crc kubenswrapper[4871]: I1007 22:56:45.839243 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9szb" event={"ID":"847482fb-1339-4a87-9d4d-f178aaf0435f","Type":"ContainerStarted","Data":"251a8fef73a4dd85035af817c7274df0bcd6a6b262e6a3574b0a8bd61c5c40d5"} Oct 07 22:56:45 crc kubenswrapper[4871]: I1007 22:56:45.841261 4871 generic.go:334] "Generic (PLEG): container finished" podID="8f7a029e-6cb3-4051-b2b2-a393044a2ea3" containerID="a43e563624a4147557840d5e957e2a3b43034fa657baa03aaee62713c7628bb9" exitCode=0 Oct 07 22:56:45 crc kubenswrapper[4871]: I1007 22:56:45.841307 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lggxl" event={"ID":"8f7a029e-6cb3-4051-b2b2-a393044a2ea3","Type":"ContainerDied","Data":"a43e563624a4147557840d5e957e2a3b43034fa657baa03aaee62713c7628bb9"} Oct 07 22:56:45 crc kubenswrapper[4871]: I1007 22:56:45.841346 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lggxl" event={"ID":"8f7a029e-6cb3-4051-b2b2-a393044a2ea3","Type":"ContainerStarted","Data":"ea2af9dd3664e61a52b3aba336c0b66e33b16a1f7e42dd1a574b4b6cfcb80870"} Oct 07 22:56:46 crc kubenswrapper[4871]: I1007 22:56:46.852514 4871 generic.go:334] "Generic (PLEG): container finished" podID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerID="f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484" exitCode=0 Oct 07 22:56:46 crc kubenswrapper[4871]: I1007 22:56:46.852566 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2tz6" event={"ID":"ba306cf0-f483-4d8d-9712-863ef4db5773","Type":"ContainerDied","Data":"f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484"} Oct 07 22:56:46 crc kubenswrapper[4871]: I1007 22:56:46.857460 4871 generic.go:334] "Generic (PLEG): container finished" podID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerID="251a8fef73a4dd85035af817c7274df0bcd6a6b262e6a3574b0a8bd61c5c40d5" exitCode=0 Oct 07 22:56:46 crc kubenswrapper[4871]: I1007 22:56:46.857501 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9szb" event={"ID":"847482fb-1339-4a87-9d4d-f178aaf0435f","Type":"ContainerDied","Data":"251a8fef73a4dd85035af817c7274df0bcd6a6b262e6a3574b0a8bd61c5c40d5"} Oct 07 22:56:48 crc kubenswrapper[4871]: I1007 22:56:48.883114 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2tz6" event={"ID":"ba306cf0-f483-4d8d-9712-863ef4db5773","Type":"ContainerStarted","Data":"599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d"} Oct 07 22:56:48 crc kubenswrapper[4871]: I1007 22:56:48.888688 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9szb" event={"ID":"847482fb-1339-4a87-9d4d-f178aaf0435f","Type":"ContainerStarted","Data":"0d93373d0261c75f315a83094994ca447a5eb806140719cd9fa85941e5c9342f"} Oct 07 22:56:48 crc kubenswrapper[4871]: I1007 22:56:48.891898 4871 generic.go:334] "Generic (PLEG): container finished" podID="8f7a029e-6cb3-4051-b2b2-a393044a2ea3" containerID="532a8ad5c4a2db077d1b77307673356478e2ab8db5b2b6a3260203bbf721ffad" exitCode=0 Oct 07 22:56:48 crc kubenswrapper[4871]: I1007 22:56:48.891994 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lggxl" event={"ID":"8f7a029e-6cb3-4051-b2b2-a393044a2ea3","Type":"ContainerDied","Data":"532a8ad5c4a2db077d1b77307673356478e2ab8db5b2b6a3260203bbf721ffad"} Oct 07 22:56:48 crc kubenswrapper[4871]: I1007 22:56:48.919834 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l2tz6" podStartSLOduration=2.312386082 podStartE2EDuration="6.919784843s" podCreationTimestamp="2025-10-07 22:56:42 +0000 UTC" firstStartedPulling="2025-10-07 22:56:43.808813315 +0000 UTC m=+2877.611511388" lastFinishedPulling="2025-10-07 22:56:48.416212066 +0000 UTC m=+2882.218910149" observedRunningTime="2025-10-07 22:56:48.916342642 +0000 UTC m=+2882.719040745" watchObservedRunningTime="2025-10-07 22:56:48.919784843 +0000 UTC m=+2882.722482936" Oct 07 22:56:48 crc kubenswrapper[4871]: I1007 22:56:48.976582 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l9szb" podStartSLOduration=2.594775808 podStartE2EDuration="5.976561671s" podCreationTimestamp="2025-10-07 22:56:43 +0000 UTC" firstStartedPulling="2025-10-07 22:56:44.821858575 +0000 UTC m=+2878.624556688" lastFinishedPulling="2025-10-07 22:56:48.203644478 +0000 UTC m=+2882.006342551" observedRunningTime="2025-10-07 22:56:48.968505869 +0000 UTC m=+2882.771203942" watchObservedRunningTime="2025-10-07 22:56:48.976561671 +0000 UTC m=+2882.779259754" Oct 07 22:56:49 crc kubenswrapper[4871]: I1007 22:56:49.902921 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lggxl" event={"ID":"8f7a029e-6cb3-4051-b2b2-a393044a2ea3","Type":"ContainerStarted","Data":"97b8377153fcf1cd027cbb9a3d8f402ce4dc78dc9a4df19df167324fe0e2133d"} Oct 07 22:56:49 crc kubenswrapper[4871]: I1007 22:56:49.931334 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lggxl" podStartSLOduration=2.131912573 podStartE2EDuration="5.931309154s" podCreationTimestamp="2025-10-07 22:56:44 +0000 UTC" firstStartedPulling="2025-10-07 22:56:45.843303717 +0000 UTC m=+2879.646001820" lastFinishedPulling="2025-10-07 22:56:49.642700328 +0000 UTC m=+2883.445398401" observedRunningTime="2025-10-07 22:56:49.927198985 +0000 UTC m=+2883.729897058" watchObservedRunningTime="2025-10-07 22:56:49.931309154 +0000 UTC m=+2883.734007257" Oct 07 22:56:52 crc kubenswrapper[4871]: I1007 22:56:52.658851 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:52 crc kubenswrapper[4871]: I1007 22:56:52.659536 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:56:53 crc kubenswrapper[4871]: I1007 22:56:53.638300 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:53 crc kubenswrapper[4871]: I1007 22:56:53.638384 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:53 crc kubenswrapper[4871]: I1007 22:56:53.699485 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:53 crc kubenswrapper[4871]: I1007 22:56:53.717610 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l2tz6" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerName="registry-server" probeResult="failure" output=< Oct 07 22:56:53 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 22:56:53 crc kubenswrapper[4871]: > Oct 07 22:56:54 crc kubenswrapper[4871]: I1007 22:56:54.015855 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:54 crc kubenswrapper[4871]: I1007 22:56:54.698724 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9szb"] Oct 07 22:56:55 crc kubenswrapper[4871]: I1007 22:56:55.062884 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:55 crc kubenswrapper[4871]: I1007 22:56:55.063227 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:55 crc kubenswrapper[4871]: I1007 22:56:55.123337 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:55 crc kubenswrapper[4871]: I1007 22:56:55.962404 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l9szb" podUID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerName="registry-server" containerID="cri-o://0d93373d0261c75f315a83094994ca447a5eb806140719cd9fa85941e5c9342f" gracePeriod=2 Oct 07 22:56:56 crc kubenswrapper[4871]: I1007 22:56:56.038977 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lggxl" Oct 07 22:56:56 crc kubenswrapper[4871]: E1007 22:56:56.218953 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod847482fb_1339_4a87_9d4d_f178aaf0435f.slice/crio-0d93373d0261c75f315a83094994ca447a5eb806140719cd9fa85941e5c9342f.scope\": RecentStats: unable to find data in memory cache]" Oct 07 22:56:56 crc kubenswrapper[4871]: I1007 22:56:56.975244 4871 generic.go:334] "Generic (PLEG): container finished" podID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerID="0d93373d0261c75f315a83094994ca447a5eb806140719cd9fa85941e5c9342f" exitCode=0 Oct 07 22:56:56 crc kubenswrapper[4871]: I1007 22:56:56.975301 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9szb" event={"ID":"847482fb-1339-4a87-9d4d-f178aaf0435f","Type":"ContainerDied","Data":"0d93373d0261c75f315a83094994ca447a5eb806140719cd9fa85941e5c9342f"} Oct 07 22:56:56 crc kubenswrapper[4871]: I1007 22:56:56.975842 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9szb" event={"ID":"847482fb-1339-4a87-9d4d-f178aaf0435f","Type":"ContainerDied","Data":"04d2c733a2c2b3ce19e3d9d4411873a9a9567d6e1c8e35272808966b23f61e25"} Oct 07 22:56:56 crc kubenswrapper[4871]: I1007 22:56:56.975868 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04d2c733a2c2b3ce19e3d9d4411873a9a9567d6e1c8e35272808966b23f61e25" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.003580 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.146169 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-utilities\") pod \"847482fb-1339-4a87-9d4d-f178aaf0435f\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.146253 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2vzx\" (UniqueName: \"kubernetes.io/projected/847482fb-1339-4a87-9d4d-f178aaf0435f-kube-api-access-z2vzx\") pod \"847482fb-1339-4a87-9d4d-f178aaf0435f\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.146373 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-catalog-content\") pod \"847482fb-1339-4a87-9d4d-f178aaf0435f\" (UID: \"847482fb-1339-4a87-9d4d-f178aaf0435f\") " Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.150990 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-utilities" (OuterVolumeSpecName: "utilities") pod "847482fb-1339-4a87-9d4d-f178aaf0435f" (UID: "847482fb-1339-4a87-9d4d-f178aaf0435f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.156039 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/847482fb-1339-4a87-9d4d-f178aaf0435f-kube-api-access-z2vzx" (OuterVolumeSpecName: "kube-api-access-z2vzx") pod "847482fb-1339-4a87-9d4d-f178aaf0435f" (UID: "847482fb-1339-4a87-9d4d-f178aaf0435f"). InnerVolumeSpecName "kube-api-access-z2vzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.227658 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lggxl"] Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.235395 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "847482fb-1339-4a87-9d4d-f178aaf0435f" (UID: "847482fb-1339-4a87-9d4d-f178aaf0435f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.250481 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.250527 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/847482fb-1339-4a87-9d4d-f178aaf0435f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.250547 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2vzx\" (UniqueName: \"kubernetes.io/projected/847482fb-1339-4a87-9d4d-f178aaf0435f-kube-api-access-z2vzx\") on node \"crc\" DevicePath \"\"" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.511232 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4jd9w"] Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.511577 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4jd9w" podUID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerName="registry-server" containerID="cri-o://9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e" gracePeriod=2 Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.952128 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.961454 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-catalog-content\") pod \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.961507 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f76dw\" (UniqueName: \"kubernetes.io/projected/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-kube-api-access-f76dw\") pod \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.961569 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-utilities\") pod \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\" (UID: \"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2\") " Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.962114 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-utilities" (OuterVolumeSpecName: "utilities") pod "6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" (UID: "6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.967458 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-kube-api-access-f76dw" (OuterVolumeSpecName: "kube-api-access-f76dw") pod "6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" (UID: "6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2"). InnerVolumeSpecName "kube-api-access-f76dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.979587 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" (UID: "6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.984878 4871 generic.go:334] "Generic (PLEG): container finished" podID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerID="9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e" exitCode=0 Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.984951 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4jd9w" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.984987 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4jd9w" event={"ID":"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2","Type":"ContainerDied","Data":"9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e"} Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.985014 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4jd9w" event={"ID":"6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2","Type":"ContainerDied","Data":"339cbec5f04200f415d95f2c07a6f30e687fb298e20aac3961d71206f2f1141f"} Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.985033 4871 scope.go:117] "RemoveContainer" containerID="9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e" Oct 07 22:56:57 crc kubenswrapper[4871]: I1007 22:56:57.985160 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9szb" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.020939 4871 scope.go:117] "RemoveContainer" containerID="7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.023747 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4jd9w"] Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.028067 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4jd9w"] Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.040051 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9szb"] Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.044875 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l9szb"] Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.051236 4871 scope.go:117] "RemoveContainer" containerID="050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.064617 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.064659 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.064670 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f76dw\" (UniqueName: \"kubernetes.io/projected/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2-kube-api-access-f76dw\") on node \"crc\" DevicePath \"\"" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.065050 4871 scope.go:117] "RemoveContainer" containerID="9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e" Oct 07 22:56:58 crc kubenswrapper[4871]: E1007 22:56:58.065463 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e\": container with ID starting with 9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e not found: ID does not exist" containerID="9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.065527 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e"} err="failed to get container status \"9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e\": rpc error: code = NotFound desc = could not find container \"9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e\": container with ID starting with 9759e1c30eda3777407a3fe8489d44dac614bc7f39a706923802a129d5ba819e not found: ID does not exist" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.065559 4871 scope.go:117] "RemoveContainer" containerID="7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f" Oct 07 22:56:58 crc kubenswrapper[4871]: E1007 22:56:58.065882 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f\": container with ID starting with 7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f not found: ID does not exist" containerID="7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.065919 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f"} err="failed to get container status \"7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f\": rpc error: code = NotFound desc = could not find container \"7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f\": container with ID starting with 7c3c8a76b008bde11274bbfe0f149811ec4ffd1be930f0a62463b5c3d3032f0f not found: ID does not exist" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.065942 4871 scope.go:117] "RemoveContainer" containerID="050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775" Oct 07 22:56:58 crc kubenswrapper[4871]: E1007 22:56:58.066208 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775\": container with ID starting with 050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775 not found: ID does not exist" containerID="050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.066249 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775"} err="failed to get container status \"050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775\": rpc error: code = NotFound desc = could not find container \"050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775\": container with ID starting with 050355cd566970bf263bb9b240d6722e812e03f4d80ea3821b539543f0c92775 not found: ID does not exist" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.991412 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" path="/var/lib/kubelet/pods/6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2/volumes" Oct 07 22:56:58 crc kubenswrapper[4871]: I1007 22:56:58.992069 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="847482fb-1339-4a87-9d4d-f178aaf0435f" path="/var/lib/kubelet/pods/847482fb-1339-4a87-9d4d-f178aaf0435f/volumes" Oct 07 22:57:02 crc kubenswrapper[4871]: I1007 22:57:02.739006 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:57:02 crc kubenswrapper[4871]: I1007 22:57:02.794701 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.296824 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l2tz6"] Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.297113 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l2tz6" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerName="registry-server" containerID="cri-o://599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d" gracePeriod=2 Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.551505 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.551569 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.777652 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.881967 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-catalog-content\") pod \"ba306cf0-f483-4d8d-9712-863ef4db5773\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.882062 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-utilities\") pod \"ba306cf0-f483-4d8d-9712-863ef4db5773\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.882111 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xqjg\" (UniqueName: \"kubernetes.io/projected/ba306cf0-f483-4d8d-9712-863ef4db5773-kube-api-access-6xqjg\") pod \"ba306cf0-f483-4d8d-9712-863ef4db5773\" (UID: \"ba306cf0-f483-4d8d-9712-863ef4db5773\") " Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.883163 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-utilities" (OuterVolumeSpecName: "utilities") pod "ba306cf0-f483-4d8d-9712-863ef4db5773" (UID: "ba306cf0-f483-4d8d-9712-863ef4db5773"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.891885 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba306cf0-f483-4d8d-9712-863ef4db5773-kube-api-access-6xqjg" (OuterVolumeSpecName: "kube-api-access-6xqjg") pod "ba306cf0-f483-4d8d-9712-863ef4db5773" (UID: "ba306cf0-f483-4d8d-9712-863ef4db5773"). InnerVolumeSpecName "kube-api-access-6xqjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.975075 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba306cf0-f483-4d8d-9712-863ef4db5773" (UID: "ba306cf0-f483-4d8d-9712-863ef4db5773"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.984857 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xqjg\" (UniqueName: \"kubernetes.io/projected/ba306cf0-f483-4d8d-9712-863ef4db5773-kube-api-access-6xqjg\") on node \"crc\" DevicePath \"\"" Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.984892 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 22:57:05 crc kubenswrapper[4871]: I1007 22:57:05.984902 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba306cf0-f483-4d8d-9712-863ef4db5773-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.079643 4871 generic.go:334] "Generic (PLEG): container finished" podID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerID="599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d" exitCode=0 Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.079692 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2tz6" event={"ID":"ba306cf0-f483-4d8d-9712-863ef4db5773","Type":"ContainerDied","Data":"599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d"} Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.079722 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2tz6" event={"ID":"ba306cf0-f483-4d8d-9712-863ef4db5773","Type":"ContainerDied","Data":"dc34d63e32aa3dad8c5a034e2d12853d10b09d0883101e6646f401821b60c89d"} Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.079744 4871 scope.go:117] "RemoveContainer" containerID="599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.079930 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2tz6" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.099177 4871 scope.go:117] "RemoveContainer" containerID="f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.124946 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l2tz6"] Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.129349 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l2tz6"] Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.131323 4871 scope.go:117] "RemoveContainer" containerID="61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.161774 4871 scope.go:117] "RemoveContainer" containerID="599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d" Oct 07 22:57:06 crc kubenswrapper[4871]: E1007 22:57:06.162220 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d\": container with ID starting with 599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d not found: ID does not exist" containerID="599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.162254 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d"} err="failed to get container status \"599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d\": rpc error: code = NotFound desc = could not find container \"599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d\": container with ID starting with 599dae3c84d62f74011bcf4386a14f78b15d0093bbcc294c7eb2295346e0a43d not found: ID does not exist" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.162278 4871 scope.go:117] "RemoveContainer" containerID="f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484" Oct 07 22:57:06 crc kubenswrapper[4871]: E1007 22:57:06.162574 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484\": container with ID starting with f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484 not found: ID does not exist" containerID="f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.162628 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484"} err="failed to get container status \"f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484\": rpc error: code = NotFound desc = could not find container \"f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484\": container with ID starting with f951fde3f99b90dff4c4bb98293c935c555eb58766ea9150e0cf6bab5bd62484 not found: ID does not exist" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.162651 4871 scope.go:117] "RemoveContainer" containerID="61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc" Oct 07 22:57:06 crc kubenswrapper[4871]: E1007 22:57:06.162925 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc\": container with ID starting with 61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc not found: ID does not exist" containerID="61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.162944 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc"} err="failed to get container status \"61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc\": rpc error: code = NotFound desc = could not find container \"61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc\": container with ID starting with 61e383e2d6808ea40e76794d34b2a59c945e13d84ac68db73f1389eed488b4fc not found: ID does not exist" Oct 07 22:57:06 crc kubenswrapper[4871]: I1007 22:57:06.993707 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" path="/var/lib/kubelet/pods/ba306cf0-f483-4d8d-9712-863ef4db5773/volumes" Oct 07 22:57:35 crc kubenswrapper[4871]: I1007 22:57:35.511910 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 22:57:35 crc kubenswrapper[4871]: I1007 22:57:35.512611 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 22:57:35 crc kubenswrapper[4871]: I1007 22:57:35.512681 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 22:57:35 crc kubenswrapper[4871]: I1007 22:57:35.513743 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 22:57:35 crc kubenswrapper[4871]: I1007 22:57:35.513853 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" gracePeriod=600 Oct 07 22:57:36 crc kubenswrapper[4871]: E1007 22:57:36.146497 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:57:36 crc kubenswrapper[4871]: I1007 22:57:36.343533 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" exitCode=0 Oct 07 22:57:36 crc kubenswrapper[4871]: I1007 22:57:36.343596 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb"} Oct 07 22:57:36 crc kubenswrapper[4871]: I1007 22:57:36.343650 4871 scope.go:117] "RemoveContainer" containerID="7ed4da1e7ab23d6ef0965429501276019d9fa05f06889389751d667e20f40e45" Oct 07 22:57:36 crc kubenswrapper[4871]: I1007 22:57:36.344531 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:57:36 crc kubenswrapper[4871]: E1007 22:57:36.344939 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:57:47 crc kubenswrapper[4871]: I1007 22:57:47.982536 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:57:47 crc kubenswrapper[4871]: E1007 22:57:47.983680 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:57:59 crc kubenswrapper[4871]: I1007 22:57:59.984622 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:57:59 crc kubenswrapper[4871]: E1007 22:57:59.986179 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:58:13 crc kubenswrapper[4871]: I1007 22:58:13.982401 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:58:13 crc kubenswrapper[4871]: E1007 22:58:13.983546 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:58:26 crc kubenswrapper[4871]: I1007 22:58:26.989862 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:58:26 crc kubenswrapper[4871]: E1007 22:58:26.990925 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:58:39 crc kubenswrapper[4871]: I1007 22:58:39.982192 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:58:39 crc kubenswrapper[4871]: E1007 22:58:39.983333 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:58:53 crc kubenswrapper[4871]: I1007 22:58:53.982603 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:58:53 crc kubenswrapper[4871]: E1007 22:58:53.984002 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:59:07 crc kubenswrapper[4871]: I1007 22:59:07.982734 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:59:07 crc kubenswrapper[4871]: E1007 22:59:07.983698 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:59:21 crc kubenswrapper[4871]: I1007 22:59:21.982157 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:59:21 crc kubenswrapper[4871]: E1007 22:59:21.983344 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:59:36 crc kubenswrapper[4871]: I1007 22:59:36.989307 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:59:36 crc kubenswrapper[4871]: E1007 22:59:36.990311 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 22:59:48 crc kubenswrapper[4871]: I1007 22:59:48.982771 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 22:59:48 crc kubenswrapper[4871]: E1007 22:59:48.983960 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.236347 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww"] Oct 07 23:00:00 crc kubenswrapper[4871]: E1007 23:00:00.241045 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerName="extract-content" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.241077 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerName="extract-content" Oct 07 23:00:00 crc kubenswrapper[4871]: E1007 23:00:00.241104 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerName="registry-server" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.241169 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerName="registry-server" Oct 07 23:00:00 crc kubenswrapper[4871]: E1007 23:00:00.241188 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerName="extract-content" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.241195 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerName="extract-content" Oct 07 23:00:00 crc kubenswrapper[4871]: E1007 23:00:00.241208 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerName="extract-utilities" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.241214 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerName="extract-utilities" Oct 07 23:00:00 crc kubenswrapper[4871]: E1007 23:00:00.241238 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerName="extract-utilities" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.241244 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerName="extract-utilities" Oct 07 23:00:00 crc kubenswrapper[4871]: E1007 23:00:00.241261 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerName="registry-server" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.241268 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerName="registry-server" Oct 07 23:00:00 crc kubenswrapper[4871]: E1007 23:00:00.241304 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerName="extract-content" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.241310 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerName="extract-content" Oct 07 23:00:00 crc kubenswrapper[4871]: E1007 23:00:00.241333 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerName="registry-server" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.241339 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerName="registry-server" Oct 07 23:00:00 crc kubenswrapper[4871]: E1007 23:00:00.241355 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerName="extract-utilities" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.241362 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerName="extract-utilities" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.244229 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="847482fb-1339-4a87-9d4d-f178aaf0435f" containerName="registry-server" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.244259 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c3eefd2-0d10-4f32-86e5-f6d7af4a2bf2" containerName="registry-server" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.244278 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba306cf0-f483-4d8d-9712-863ef4db5773" containerName="registry-server" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.245942 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.249738 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.250021 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.274914 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww"] Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.379128 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zdv4\" (UniqueName: \"kubernetes.io/projected/034e49d6-9455-40d9-98fa-83d7b737a3c0-kube-api-access-9zdv4\") pod \"collect-profiles-29331300-4llww\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.379201 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034e49d6-9455-40d9-98fa-83d7b737a3c0-config-volume\") pod \"collect-profiles-29331300-4llww\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.379532 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034e49d6-9455-40d9-98fa-83d7b737a3c0-secret-volume\") pod \"collect-profiles-29331300-4llww\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.481333 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034e49d6-9455-40d9-98fa-83d7b737a3c0-secret-volume\") pod \"collect-profiles-29331300-4llww\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.481713 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zdv4\" (UniqueName: \"kubernetes.io/projected/034e49d6-9455-40d9-98fa-83d7b737a3c0-kube-api-access-9zdv4\") pod \"collect-profiles-29331300-4llww\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.481864 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034e49d6-9455-40d9-98fa-83d7b737a3c0-config-volume\") pod \"collect-profiles-29331300-4llww\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.483932 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034e49d6-9455-40d9-98fa-83d7b737a3c0-config-volume\") pod \"collect-profiles-29331300-4llww\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.492698 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034e49d6-9455-40d9-98fa-83d7b737a3c0-secret-volume\") pod \"collect-profiles-29331300-4llww\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.519476 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zdv4\" (UniqueName: \"kubernetes.io/projected/034e49d6-9455-40d9-98fa-83d7b737a3c0-kube-api-access-9zdv4\") pod \"collect-profiles-29331300-4llww\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:00 crc kubenswrapper[4871]: I1007 23:00:00.573108 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:01 crc kubenswrapper[4871]: I1007 23:00:01.108389 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww"] Oct 07 23:00:01 crc kubenswrapper[4871]: I1007 23:00:01.712662 4871 generic.go:334] "Generic (PLEG): container finished" podID="034e49d6-9455-40d9-98fa-83d7b737a3c0" containerID="3f8117e13ab5e1f0122f7cdaa0e7186308a0aa94c58dc8efbcc0f2746b40cc30" exitCode=0 Oct 07 23:00:01 crc kubenswrapper[4871]: I1007 23:00:01.712719 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" event={"ID":"034e49d6-9455-40d9-98fa-83d7b737a3c0","Type":"ContainerDied","Data":"3f8117e13ab5e1f0122f7cdaa0e7186308a0aa94c58dc8efbcc0f2746b40cc30"} Oct 07 23:00:01 crc kubenswrapper[4871]: I1007 23:00:01.712743 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" event={"ID":"034e49d6-9455-40d9-98fa-83d7b737a3c0","Type":"ContainerStarted","Data":"420c34a79340ab5ccd781569bc013d1c142d6616c2c50021c668f5e1806cd9d1"} Oct 07 23:00:02 crc kubenswrapper[4871]: I1007 23:00:02.982292 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:00:02 crc kubenswrapper[4871]: E1007 23:00:02.983112 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.103975 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.231192 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zdv4\" (UniqueName: \"kubernetes.io/projected/034e49d6-9455-40d9-98fa-83d7b737a3c0-kube-api-access-9zdv4\") pod \"034e49d6-9455-40d9-98fa-83d7b737a3c0\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.231326 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034e49d6-9455-40d9-98fa-83d7b737a3c0-config-volume\") pod \"034e49d6-9455-40d9-98fa-83d7b737a3c0\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.231576 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034e49d6-9455-40d9-98fa-83d7b737a3c0-secret-volume\") pod \"034e49d6-9455-40d9-98fa-83d7b737a3c0\" (UID: \"034e49d6-9455-40d9-98fa-83d7b737a3c0\") " Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.232754 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/034e49d6-9455-40d9-98fa-83d7b737a3c0-config-volume" (OuterVolumeSpecName: "config-volume") pod "034e49d6-9455-40d9-98fa-83d7b737a3c0" (UID: "034e49d6-9455-40d9-98fa-83d7b737a3c0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.240926 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034e49d6-9455-40d9-98fa-83d7b737a3c0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "034e49d6-9455-40d9-98fa-83d7b737a3c0" (UID: "034e49d6-9455-40d9-98fa-83d7b737a3c0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.240979 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/034e49d6-9455-40d9-98fa-83d7b737a3c0-kube-api-access-9zdv4" (OuterVolumeSpecName: "kube-api-access-9zdv4") pod "034e49d6-9455-40d9-98fa-83d7b737a3c0" (UID: "034e49d6-9455-40d9-98fa-83d7b737a3c0"). InnerVolumeSpecName "kube-api-access-9zdv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.333752 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/034e49d6-9455-40d9-98fa-83d7b737a3c0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.333892 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zdv4\" (UniqueName: \"kubernetes.io/projected/034e49d6-9455-40d9-98fa-83d7b737a3c0-kube-api-access-9zdv4\") on node \"crc\" DevicePath \"\"" Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.333924 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/034e49d6-9455-40d9-98fa-83d7b737a3c0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.732065 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" event={"ID":"034e49d6-9455-40d9-98fa-83d7b737a3c0","Type":"ContainerDied","Data":"420c34a79340ab5ccd781569bc013d1c142d6616c2c50021c668f5e1806cd9d1"} Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.732108 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="420c34a79340ab5ccd781569bc013d1c142d6616c2c50021c668f5e1806cd9d1" Oct 07 23:00:03 crc kubenswrapper[4871]: I1007 23:00:03.732113 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww" Oct 07 23:00:04 crc kubenswrapper[4871]: I1007 23:00:04.203678 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp"] Oct 07 23:00:04 crc kubenswrapper[4871]: I1007 23:00:04.213937 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331255-cb9hp"] Oct 07 23:00:04 crc kubenswrapper[4871]: I1007 23:00:04.999745 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e7542f8-e26a-48e6-ba07-655b8d63b458" path="/var/lib/kubelet/pods/0e7542f8-e26a-48e6-ba07-655b8d63b458/volumes" Oct 07 23:00:15 crc kubenswrapper[4871]: I1007 23:00:15.982935 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:00:15 crc kubenswrapper[4871]: E1007 23:00:15.983952 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:00:27 crc kubenswrapper[4871]: I1007 23:00:27.982565 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:00:27 crc kubenswrapper[4871]: E1007 23:00:27.983569 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:00:42 crc kubenswrapper[4871]: I1007 23:00:42.983178 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:00:42 crc kubenswrapper[4871]: E1007 23:00:42.985684 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:00:54 crc kubenswrapper[4871]: I1007 23:00:54.829136 4871 scope.go:117] "RemoveContainer" containerID="871e2f33d5b297daf81fb62d40b9653b0696c86746858e34f8908e0ea9d4d33d" Oct 07 23:00:55 crc kubenswrapper[4871]: I1007 23:00:55.982190 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:00:55 crc kubenswrapper[4871]: E1007 23:00:55.983142 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:01:09 crc kubenswrapper[4871]: I1007 23:01:09.982138 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:01:09 crc kubenswrapper[4871]: E1007 23:01:09.983245 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:01:22 crc kubenswrapper[4871]: I1007 23:01:22.982297 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:01:22 crc kubenswrapper[4871]: E1007 23:01:22.983261 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:01:37 crc kubenswrapper[4871]: I1007 23:01:37.982781 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:01:37 crc kubenswrapper[4871]: E1007 23:01:37.983254 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:01:49 crc kubenswrapper[4871]: I1007 23:01:49.982285 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:01:49 crc kubenswrapper[4871]: E1007 23:01:49.983417 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:02:01 crc kubenswrapper[4871]: I1007 23:02:01.982080 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:02:01 crc kubenswrapper[4871]: E1007 23:02:01.982989 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:02:14 crc kubenswrapper[4871]: I1007 23:02:14.982827 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:02:14 crc kubenswrapper[4871]: E1007 23:02:14.983845 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:02:28 crc kubenswrapper[4871]: I1007 23:02:28.983273 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:02:28 crc kubenswrapper[4871]: E1007 23:02:28.984254 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:02:43 crc kubenswrapper[4871]: I1007 23:02:43.983221 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:02:44 crc kubenswrapper[4871]: I1007 23:02:44.252008 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"4b53bd460ead0087e0d627e65163948ecfc09a4a5d94d569be0e1eeaab97ec3d"} Oct 07 23:02:54 crc kubenswrapper[4871]: I1007 23:02:54.915577 4871 scope.go:117] "RemoveContainer" containerID="0d93373d0261c75f315a83094994ca447a5eb806140719cd9fa85941e5c9342f" Oct 07 23:02:54 crc kubenswrapper[4871]: I1007 23:02:54.947871 4871 scope.go:117] "RemoveContainer" containerID="251a8fef73a4dd85035af817c7274df0bcd6a6b262e6a3574b0a8bd61c5c40d5" Oct 07 23:02:54 crc kubenswrapper[4871]: I1007 23:02:54.993448 4871 scope.go:117] "RemoveContainer" containerID="56793ceef6a6908856d3879a985a2163e075bb1e0dd6d3328e8119ff4172e4e4" Oct 07 23:05:05 crc kubenswrapper[4871]: I1007 23:05:05.512717 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:05:05 crc kubenswrapper[4871]: I1007 23:05:05.513276 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:05:35 crc kubenswrapper[4871]: I1007 23:05:35.512206 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:05:35 crc kubenswrapper[4871]: I1007 23:05:35.512972 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:06:05 crc kubenswrapper[4871]: I1007 23:06:05.512693 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:06:05 crc kubenswrapper[4871]: I1007 23:06:05.513571 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:06:05 crc kubenswrapper[4871]: I1007 23:06:05.513646 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:06:05 crc kubenswrapper[4871]: I1007 23:06:05.514863 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b53bd460ead0087e0d627e65163948ecfc09a4a5d94d569be0e1eeaab97ec3d"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:06:05 crc kubenswrapper[4871]: I1007 23:06:05.514989 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://4b53bd460ead0087e0d627e65163948ecfc09a4a5d94d569be0e1eeaab97ec3d" gracePeriod=600 Oct 07 23:06:06 crc kubenswrapper[4871]: I1007 23:06:06.131685 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="4b53bd460ead0087e0d627e65163948ecfc09a4a5d94d569be0e1eeaab97ec3d" exitCode=0 Oct 07 23:06:06 crc kubenswrapper[4871]: I1007 23:06:06.131755 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"4b53bd460ead0087e0d627e65163948ecfc09a4a5d94d569be0e1eeaab97ec3d"} Oct 07 23:06:06 crc kubenswrapper[4871]: I1007 23:06:06.132019 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a"} Oct 07 23:06:06 crc kubenswrapper[4871]: I1007 23:06:06.132135 4871 scope.go:117] "RemoveContainer" containerID="e6715d38e8e268d0a12256f0b47df955807e62c78310ff77b9c0745942d15abb" Oct 07 23:06:46 crc kubenswrapper[4871]: I1007 23:06:46.968542 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hxf2w"] Oct 07 23:06:46 crc kubenswrapper[4871]: E1007 23:06:46.969295 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034e49d6-9455-40d9-98fa-83d7b737a3c0" containerName="collect-profiles" Oct 07 23:06:46 crc kubenswrapper[4871]: I1007 23:06:46.969306 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="034e49d6-9455-40d9-98fa-83d7b737a3c0" containerName="collect-profiles" Oct 07 23:06:46 crc kubenswrapper[4871]: I1007 23:06:46.969459 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="034e49d6-9455-40d9-98fa-83d7b737a3c0" containerName="collect-profiles" Oct 07 23:06:46 crc kubenswrapper[4871]: I1007 23:06:46.970639 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:46.999287 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-catalog-content\") pod \"community-operators-hxf2w\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:46.999852 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxgns\" (UniqueName: \"kubernetes.io/projected/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-kube-api-access-bxgns\") pod \"community-operators-hxf2w\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:46.999904 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-utilities\") pod \"community-operators-hxf2w\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:47.009981 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hxf2w"] Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:47.101312 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-utilities\") pod \"community-operators-hxf2w\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:47.101391 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-catalog-content\") pod \"community-operators-hxf2w\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:47.101448 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxgns\" (UniqueName: \"kubernetes.io/projected/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-kube-api-access-bxgns\") pod \"community-operators-hxf2w\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:47.102140 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-utilities\") pod \"community-operators-hxf2w\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:47.102377 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-catalog-content\") pod \"community-operators-hxf2w\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:47.127166 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxgns\" (UniqueName: \"kubernetes.io/projected/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-kube-api-access-bxgns\") pod \"community-operators-hxf2w\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:47.300928 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:47 crc kubenswrapper[4871]: I1007 23:06:47.544177 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hxf2w"] Oct 07 23:06:48 crc kubenswrapper[4871]: I1007 23:06:48.532712 4871 generic.go:334] "Generic (PLEG): container finished" podID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerID="664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d" exitCode=0 Oct 07 23:06:48 crc kubenswrapper[4871]: I1007 23:06:48.532754 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxf2w" event={"ID":"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2","Type":"ContainerDied","Data":"664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d"} Oct 07 23:06:48 crc kubenswrapper[4871]: I1007 23:06:48.532777 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxf2w" event={"ID":"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2","Type":"ContainerStarted","Data":"e0d1afaa3a7e07184342815cb15795a005124a0dfafbbe6f173ccb7061094f6a"} Oct 07 23:06:48 crc kubenswrapper[4871]: I1007 23:06:48.536848 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 23:06:49 crc kubenswrapper[4871]: I1007 23:06:49.546087 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxf2w" event={"ID":"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2","Type":"ContainerStarted","Data":"f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4"} Oct 07 23:06:50 crc kubenswrapper[4871]: I1007 23:06:50.559058 4871 generic.go:334] "Generic (PLEG): container finished" podID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerID="f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4" exitCode=0 Oct 07 23:06:50 crc kubenswrapper[4871]: I1007 23:06:50.559137 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxf2w" event={"ID":"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2","Type":"ContainerDied","Data":"f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4"} Oct 07 23:06:51 crc kubenswrapper[4871]: I1007 23:06:51.572776 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxf2w" event={"ID":"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2","Type":"ContainerStarted","Data":"b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d"} Oct 07 23:06:51 crc kubenswrapper[4871]: I1007 23:06:51.605168 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hxf2w" podStartSLOduration=3.137479978 podStartE2EDuration="5.605128502s" podCreationTimestamp="2025-10-07 23:06:46 +0000 UTC" firstStartedPulling="2025-10-07 23:06:48.535915853 +0000 UTC m=+3482.338613956" lastFinishedPulling="2025-10-07 23:06:51.003564367 +0000 UTC m=+3484.806262480" observedRunningTime="2025-10-07 23:06:51.59902395 +0000 UTC m=+3485.401722063" watchObservedRunningTime="2025-10-07 23:06:51.605128502 +0000 UTC m=+3485.407826645" Oct 07 23:06:57 crc kubenswrapper[4871]: I1007 23:06:57.301806 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:57 crc kubenswrapper[4871]: I1007 23:06:57.303056 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:57 crc kubenswrapper[4871]: I1007 23:06:57.366059 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:57 crc kubenswrapper[4871]: I1007 23:06:57.690487 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:06:57 crc kubenswrapper[4871]: I1007 23:06:57.748658 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hxf2w"] Oct 07 23:06:59 crc kubenswrapper[4871]: I1007 23:06:59.646696 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hxf2w" podUID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerName="registry-server" containerID="cri-o://b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d" gracePeriod=2 Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.158135 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.314657 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-utilities\") pod \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.314764 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxgns\" (UniqueName: \"kubernetes.io/projected/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-kube-api-access-bxgns\") pod \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.314862 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-catalog-content\") pod \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\" (UID: \"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2\") " Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.316578 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-utilities" (OuterVolumeSpecName: "utilities") pod "68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" (UID: "68d2d636-ed3c-460c-bbdc-594f0fd2a2b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.327033 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-kube-api-access-bxgns" (OuterVolumeSpecName: "kube-api-access-bxgns") pod "68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" (UID: "68d2d636-ed3c-460c-bbdc-594f0fd2a2b2"). InnerVolumeSpecName "kube-api-access-bxgns". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.388348 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" (UID: "68d2d636-ed3c-460c-bbdc-594f0fd2a2b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.417109 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.417165 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxgns\" (UniqueName: \"kubernetes.io/projected/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-kube-api-access-bxgns\") on node \"crc\" DevicePath \"\"" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.417188 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.660694 4871 generic.go:334] "Generic (PLEG): container finished" podID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerID="b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d" exitCode=0 Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.660754 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxf2w" event={"ID":"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2","Type":"ContainerDied","Data":"b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d"} Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.660850 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxf2w" event={"ID":"68d2d636-ed3c-460c-bbdc-594f0fd2a2b2","Type":"ContainerDied","Data":"e0d1afaa3a7e07184342815cb15795a005124a0dfafbbe6f173ccb7061094f6a"} Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.660897 4871 scope.go:117] "RemoveContainer" containerID="b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.660922 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxf2w" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.689688 4871 scope.go:117] "RemoveContainer" containerID="f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.724724 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hxf2w"] Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.735155 4871 scope.go:117] "RemoveContainer" containerID="664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.736374 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hxf2w"] Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.767889 4871 scope.go:117] "RemoveContainer" containerID="b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d" Oct 07 23:07:00 crc kubenswrapper[4871]: E1007 23:07:00.768943 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d\": container with ID starting with b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d not found: ID does not exist" containerID="b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.769013 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d"} err="failed to get container status \"b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d\": rpc error: code = NotFound desc = could not find container \"b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d\": container with ID starting with b0bbc1684f75a6b13a41b38f5695eb480377df5eccc0fbdebd0f924bad36526d not found: ID does not exist" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.769054 4871 scope.go:117] "RemoveContainer" containerID="f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4" Oct 07 23:07:00 crc kubenswrapper[4871]: E1007 23:07:00.770181 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4\": container with ID starting with f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4 not found: ID does not exist" containerID="f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.770224 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4"} err="failed to get container status \"f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4\": rpc error: code = NotFound desc = could not find container \"f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4\": container with ID starting with f95cf8556983b400d81a3f924372c471ee72fe27c58b72957047a22178b12ad4 not found: ID does not exist" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.770252 4871 scope.go:117] "RemoveContainer" containerID="664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d" Oct 07 23:07:00 crc kubenswrapper[4871]: E1007 23:07:00.771062 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d\": container with ID starting with 664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d not found: ID does not exist" containerID="664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.771147 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d"} err="failed to get container status \"664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d\": rpc error: code = NotFound desc = could not find container \"664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d\": container with ID starting with 664235d3ef5bb7817d7e36b7bbe0a9d91d72b4d11f8db69a9a4b0196f4736c5d not found: ID does not exist" Oct 07 23:07:00 crc kubenswrapper[4871]: I1007 23:07:00.994931 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" path="/var/lib/kubelet/pods/68d2d636-ed3c-460c-bbdc-594f0fd2a2b2/volumes" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.032485 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nrdhx"] Oct 07 23:07:03 crc kubenswrapper[4871]: E1007 23:07:03.033328 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerName="registry-server" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.033351 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerName="registry-server" Oct 07 23:07:03 crc kubenswrapper[4871]: E1007 23:07:03.033386 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerName="extract-utilities" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.033402 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerName="extract-utilities" Oct 07 23:07:03 crc kubenswrapper[4871]: E1007 23:07:03.033433 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerName="extract-content" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.033450 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerName="extract-content" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.033699 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d2d636-ed3c-460c-bbdc-594f0fd2a2b2" containerName="registry-server" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.035764 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.041331 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrdhx"] Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.162118 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-utilities\") pod \"redhat-marketplace-nrdhx\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.162257 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbddt\" (UniqueName: \"kubernetes.io/projected/97f05706-cd79-4fd2-b936-92f4b6463647-kube-api-access-hbddt\") pod \"redhat-marketplace-nrdhx\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.162428 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-catalog-content\") pod \"redhat-marketplace-nrdhx\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.263872 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbddt\" (UniqueName: \"kubernetes.io/projected/97f05706-cd79-4fd2-b936-92f4b6463647-kube-api-access-hbddt\") pod \"redhat-marketplace-nrdhx\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.263942 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-catalog-content\") pod \"redhat-marketplace-nrdhx\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.264002 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-utilities\") pod \"redhat-marketplace-nrdhx\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.264453 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-utilities\") pod \"redhat-marketplace-nrdhx\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.264619 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-catalog-content\") pod \"redhat-marketplace-nrdhx\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.292947 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbddt\" (UniqueName: \"kubernetes.io/projected/97f05706-cd79-4fd2-b936-92f4b6463647-kube-api-access-hbddt\") pod \"redhat-marketplace-nrdhx\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.372676 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.622679 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrdhx"] Oct 07 23:07:03 crc kubenswrapper[4871]: I1007 23:07:03.688895 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrdhx" event={"ID":"97f05706-cd79-4fd2-b936-92f4b6463647","Type":"ContainerStarted","Data":"563ac7285e2a1d459abaaf1a2275675f0d16f434ba1bbabe9618f74704e8b179"} Oct 07 23:07:04 crc kubenswrapper[4871]: I1007 23:07:04.703294 4871 generic.go:334] "Generic (PLEG): container finished" podID="97f05706-cd79-4fd2-b936-92f4b6463647" containerID="c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656" exitCode=0 Oct 07 23:07:04 crc kubenswrapper[4871]: I1007 23:07:04.703403 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrdhx" event={"ID":"97f05706-cd79-4fd2-b936-92f4b6463647","Type":"ContainerDied","Data":"c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656"} Oct 07 23:07:06 crc kubenswrapper[4871]: I1007 23:07:06.727857 4871 generic.go:334] "Generic (PLEG): container finished" podID="97f05706-cd79-4fd2-b936-92f4b6463647" containerID="5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05" exitCode=0 Oct 07 23:07:06 crc kubenswrapper[4871]: I1007 23:07:06.729537 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrdhx" event={"ID":"97f05706-cd79-4fd2-b936-92f4b6463647","Type":"ContainerDied","Data":"5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05"} Oct 07 23:07:07 crc kubenswrapper[4871]: I1007 23:07:07.740881 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrdhx" event={"ID":"97f05706-cd79-4fd2-b936-92f4b6463647","Type":"ContainerStarted","Data":"999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9"} Oct 07 23:07:07 crc kubenswrapper[4871]: I1007 23:07:07.766989 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nrdhx" podStartSLOduration=2.226922563 podStartE2EDuration="4.766957653s" podCreationTimestamp="2025-10-07 23:07:03 +0000 UTC" firstStartedPulling="2025-10-07 23:07:04.708698841 +0000 UTC m=+3498.511396924" lastFinishedPulling="2025-10-07 23:07:07.248733911 +0000 UTC m=+3501.051432014" observedRunningTime="2025-10-07 23:07:07.761533289 +0000 UTC m=+3501.564231402" watchObservedRunningTime="2025-10-07 23:07:07.766957653 +0000 UTC m=+3501.569655746" Oct 07 23:07:13 crc kubenswrapper[4871]: I1007 23:07:13.373214 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:13 crc kubenswrapper[4871]: I1007 23:07:13.374032 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:13 crc kubenswrapper[4871]: I1007 23:07:13.450474 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:13 crc kubenswrapper[4871]: I1007 23:07:13.865617 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:13 crc kubenswrapper[4871]: I1007 23:07:13.926398 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrdhx"] Oct 07 23:07:15 crc kubenswrapper[4871]: I1007 23:07:15.809250 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nrdhx" podUID="97f05706-cd79-4fd2-b936-92f4b6463647" containerName="registry-server" containerID="cri-o://999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9" gracePeriod=2 Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.249363 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.364609 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-catalog-content\") pod \"97f05706-cd79-4fd2-b936-92f4b6463647\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.364785 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbddt\" (UniqueName: \"kubernetes.io/projected/97f05706-cd79-4fd2-b936-92f4b6463647-kube-api-access-hbddt\") pod \"97f05706-cd79-4fd2-b936-92f4b6463647\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.364835 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-utilities\") pod \"97f05706-cd79-4fd2-b936-92f4b6463647\" (UID: \"97f05706-cd79-4fd2-b936-92f4b6463647\") " Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.365884 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-utilities" (OuterVolumeSpecName: "utilities") pod "97f05706-cd79-4fd2-b936-92f4b6463647" (UID: "97f05706-cd79-4fd2-b936-92f4b6463647"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.371036 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97f05706-cd79-4fd2-b936-92f4b6463647-kube-api-access-hbddt" (OuterVolumeSpecName: "kube-api-access-hbddt") pod "97f05706-cd79-4fd2-b936-92f4b6463647" (UID: "97f05706-cd79-4fd2-b936-92f4b6463647"). InnerVolumeSpecName "kube-api-access-hbddt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.378193 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97f05706-cd79-4fd2-b936-92f4b6463647" (UID: "97f05706-cd79-4fd2-b936-92f4b6463647"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.466726 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbddt\" (UniqueName: \"kubernetes.io/projected/97f05706-cd79-4fd2-b936-92f4b6463647-kube-api-access-hbddt\") on node \"crc\" DevicePath \"\"" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.466822 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.466853 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f05706-cd79-4fd2-b936-92f4b6463647-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.822603 4871 generic.go:334] "Generic (PLEG): container finished" podID="97f05706-cd79-4fd2-b936-92f4b6463647" containerID="999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9" exitCode=0 Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.822662 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrdhx" event={"ID":"97f05706-cd79-4fd2-b936-92f4b6463647","Type":"ContainerDied","Data":"999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9"} Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.822706 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrdhx" event={"ID":"97f05706-cd79-4fd2-b936-92f4b6463647","Type":"ContainerDied","Data":"563ac7285e2a1d459abaaf1a2275675f0d16f434ba1bbabe9618f74704e8b179"} Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.822703 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrdhx" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.822729 4871 scope.go:117] "RemoveContainer" containerID="999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.874488 4871 scope.go:117] "RemoveContainer" containerID="5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.884888 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrdhx"] Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.895071 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrdhx"] Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.901365 4871 scope.go:117] "RemoveContainer" containerID="c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.948705 4871 scope.go:117] "RemoveContainer" containerID="999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9" Oct 07 23:07:16 crc kubenswrapper[4871]: E1007 23:07:16.949108 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9\": container with ID starting with 999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9 not found: ID does not exist" containerID="999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.949138 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9"} err="failed to get container status \"999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9\": rpc error: code = NotFound desc = could not find container \"999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9\": container with ID starting with 999a7090287fa371e085e6b9611dd1ef5f18f3347283cb4423922367549a8be9 not found: ID does not exist" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.949157 4871 scope.go:117] "RemoveContainer" containerID="5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05" Oct 07 23:07:16 crc kubenswrapper[4871]: E1007 23:07:16.949480 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05\": container with ID starting with 5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05 not found: ID does not exist" containerID="5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.949535 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05"} err="failed to get container status \"5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05\": rpc error: code = NotFound desc = could not find container \"5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05\": container with ID starting with 5f0a1d932beb1b4e8170c2c7062a89e491c2c5a4e300347f24fe2d2b3dcced05 not found: ID does not exist" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.949578 4871 scope.go:117] "RemoveContainer" containerID="c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656" Oct 07 23:07:16 crc kubenswrapper[4871]: E1007 23:07:16.949916 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656\": container with ID starting with c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656 not found: ID does not exist" containerID="c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656" Oct 07 23:07:16 crc kubenswrapper[4871]: I1007 23:07:16.949936 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656"} err="failed to get container status \"c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656\": rpc error: code = NotFound desc = could not find container \"c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656\": container with ID starting with c892499309d0b3d5d78a89a7ad2723a79b9a063c2ac59a98c594b7072f60d656 not found: ID does not exist" Oct 07 23:07:17 crc kubenswrapper[4871]: I1007 23:07:17.005667 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97f05706-cd79-4fd2-b936-92f4b6463647" path="/var/lib/kubelet/pods/97f05706-cd79-4fd2-b936-92f4b6463647/volumes" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.165518 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w8kfc"] Oct 07 23:07:27 crc kubenswrapper[4871]: E1007 23:07:27.166685 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f05706-cd79-4fd2-b936-92f4b6463647" containerName="extract-utilities" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.166708 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f05706-cd79-4fd2-b936-92f4b6463647" containerName="extract-utilities" Oct 07 23:07:27 crc kubenswrapper[4871]: E1007 23:07:27.166741 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f05706-cd79-4fd2-b936-92f4b6463647" containerName="extract-content" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.166755 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f05706-cd79-4fd2-b936-92f4b6463647" containerName="extract-content" Oct 07 23:07:27 crc kubenswrapper[4871]: E1007 23:07:27.167259 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f05706-cd79-4fd2-b936-92f4b6463647" containerName="registry-server" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.167282 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f05706-cd79-4fd2-b936-92f4b6463647" containerName="registry-server" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.167536 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="97f05706-cd79-4fd2-b936-92f4b6463647" containerName="registry-server" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.169451 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.182073 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8kfc"] Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.350647 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-utilities\") pod \"redhat-operators-w8kfc\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.351464 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-catalog-content\") pod \"redhat-operators-w8kfc\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.351581 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n29nx\" (UniqueName: \"kubernetes.io/projected/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-kube-api-access-n29nx\") pod \"redhat-operators-w8kfc\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.452772 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-catalog-content\") pod \"redhat-operators-w8kfc\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.452887 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n29nx\" (UniqueName: \"kubernetes.io/projected/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-kube-api-access-n29nx\") pod \"redhat-operators-w8kfc\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.452928 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-utilities\") pod \"redhat-operators-w8kfc\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.453444 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-catalog-content\") pod \"redhat-operators-w8kfc\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.453511 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-utilities\") pod \"redhat-operators-w8kfc\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.495016 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n29nx\" (UniqueName: \"kubernetes.io/projected/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-kube-api-access-n29nx\") pod \"redhat-operators-w8kfc\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.496418 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.930475 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8kfc"] Oct 07 23:07:27 crc kubenswrapper[4871]: I1007 23:07:27.961498 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8kfc" event={"ID":"74e5397d-d0cb-4f02-890a-e59df5fdf7fa","Type":"ContainerStarted","Data":"aacdb9ab20867bd1ca4670657f554141feb05ee134b46fb534290b7d4af2c5fd"} Oct 07 23:07:28 crc kubenswrapper[4871]: I1007 23:07:28.972848 4871 generic.go:334] "Generic (PLEG): container finished" podID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerID="11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a" exitCode=0 Oct 07 23:07:28 crc kubenswrapper[4871]: I1007 23:07:28.972985 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8kfc" event={"ID":"74e5397d-d0cb-4f02-890a-e59df5fdf7fa","Type":"ContainerDied","Data":"11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a"} Oct 07 23:07:30 crc kubenswrapper[4871]: I1007 23:07:30.994912 4871 generic.go:334] "Generic (PLEG): container finished" podID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerID="8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a" exitCode=0 Oct 07 23:07:30 crc kubenswrapper[4871]: I1007 23:07:30.996164 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8kfc" event={"ID":"74e5397d-d0cb-4f02-890a-e59df5fdf7fa","Type":"ContainerDied","Data":"8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a"} Oct 07 23:07:32 crc kubenswrapper[4871]: I1007 23:07:32.008006 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8kfc" event={"ID":"74e5397d-d0cb-4f02-890a-e59df5fdf7fa","Type":"ContainerStarted","Data":"4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9"} Oct 07 23:07:32 crc kubenswrapper[4871]: I1007 23:07:32.048168 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w8kfc" podStartSLOduration=2.563035305 podStartE2EDuration="5.048124515s" podCreationTimestamp="2025-10-07 23:07:27 +0000 UTC" firstStartedPulling="2025-10-07 23:07:28.974835224 +0000 UTC m=+3522.777533317" lastFinishedPulling="2025-10-07 23:07:31.459924414 +0000 UTC m=+3525.262622527" observedRunningTime="2025-10-07 23:07:32.039710203 +0000 UTC m=+3525.842408276" watchObservedRunningTime="2025-10-07 23:07:32.048124515 +0000 UTC m=+3525.850822588" Oct 07 23:07:37 crc kubenswrapper[4871]: I1007 23:07:37.496608 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:37 crc kubenswrapper[4871]: I1007 23:07:37.497365 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:37 crc kubenswrapper[4871]: I1007 23:07:37.558346 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:38 crc kubenswrapper[4871]: I1007 23:07:38.131496 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:38 crc kubenswrapper[4871]: I1007 23:07:38.204768 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8kfc"] Oct 07 23:07:40 crc kubenswrapper[4871]: I1007 23:07:40.075923 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w8kfc" podUID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerName="registry-server" containerID="cri-o://4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9" gracePeriod=2 Oct 07 23:07:40 crc kubenswrapper[4871]: I1007 23:07:40.512386 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:40 crc kubenswrapper[4871]: I1007 23:07:40.522873 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n29nx\" (UniqueName: \"kubernetes.io/projected/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-kube-api-access-n29nx\") pod \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " Oct 07 23:07:40 crc kubenswrapper[4871]: I1007 23:07:40.522998 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-utilities\") pod \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " Oct 07 23:07:40 crc kubenswrapper[4871]: I1007 23:07:40.523028 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-catalog-content\") pod \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\" (UID: \"74e5397d-d0cb-4f02-890a-e59df5fdf7fa\") " Oct 07 23:07:40 crc kubenswrapper[4871]: I1007 23:07:40.524124 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-utilities" (OuterVolumeSpecName: "utilities") pod "74e5397d-d0cb-4f02-890a-e59df5fdf7fa" (UID: "74e5397d-d0cb-4f02-890a-e59df5fdf7fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:07:40 crc kubenswrapper[4871]: I1007 23:07:40.529037 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-kube-api-access-n29nx" (OuterVolumeSpecName: "kube-api-access-n29nx") pod "74e5397d-d0cb-4f02-890a-e59df5fdf7fa" (UID: "74e5397d-d0cb-4f02-890a-e59df5fdf7fa"). InnerVolumeSpecName "kube-api-access-n29nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:07:40 crc kubenswrapper[4871]: I1007 23:07:40.624230 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:07:40 crc kubenswrapper[4871]: I1007 23:07:40.624287 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n29nx\" (UniqueName: \"kubernetes.io/projected/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-kube-api-access-n29nx\") on node \"crc\" DevicePath \"\"" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.084695 4871 generic.go:334] "Generic (PLEG): container finished" podID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerID="4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9" exitCode=0 Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.084740 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8kfc" event={"ID":"74e5397d-d0cb-4f02-890a-e59df5fdf7fa","Type":"ContainerDied","Data":"4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9"} Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.084769 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8kfc" event={"ID":"74e5397d-d0cb-4f02-890a-e59df5fdf7fa","Type":"ContainerDied","Data":"aacdb9ab20867bd1ca4670657f554141feb05ee134b46fb534290b7d4af2c5fd"} Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.084778 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8kfc" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.084815 4871 scope.go:117] "RemoveContainer" containerID="4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.107139 4871 scope.go:117] "RemoveContainer" containerID="8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.130875 4871 scope.go:117] "RemoveContainer" containerID="11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.158234 4871 scope.go:117] "RemoveContainer" containerID="4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9" Oct 07 23:07:41 crc kubenswrapper[4871]: E1007 23:07:41.158742 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9\": container with ID starting with 4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9 not found: ID does not exist" containerID="4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.158783 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9"} err="failed to get container status \"4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9\": rpc error: code = NotFound desc = could not find container \"4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9\": container with ID starting with 4d140d7e880911a74044973d49ed3ee05ecf6b54a886a1ec31ec562b0d90c5f9 not found: ID does not exist" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.158831 4871 scope.go:117] "RemoveContainer" containerID="8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a" Oct 07 23:07:41 crc kubenswrapper[4871]: E1007 23:07:41.159315 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a\": container with ID starting with 8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a not found: ID does not exist" containerID="8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.159370 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a"} err="failed to get container status \"8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a\": rpc error: code = NotFound desc = could not find container \"8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a\": container with ID starting with 8111ae87f5503b8cb3f64fee080a8e1ad7593871be54d82895480a202ab6c78a not found: ID does not exist" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.159401 4871 scope.go:117] "RemoveContainer" containerID="11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a" Oct 07 23:07:41 crc kubenswrapper[4871]: E1007 23:07:41.159771 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a\": container with ID starting with 11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a not found: ID does not exist" containerID="11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a" Oct 07 23:07:41 crc kubenswrapper[4871]: I1007 23:07:41.159814 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a"} err="failed to get container status \"11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a\": rpc error: code = NotFound desc = could not find container \"11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a\": container with ID starting with 11d1e46d6878f7566ba081e836ef36f7cce4eb94d96b801e8883a73257a9c63a not found: ID does not exist" Oct 07 23:07:42 crc kubenswrapper[4871]: I1007 23:07:42.502483 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74e5397d-d0cb-4f02-890a-e59df5fdf7fa" (UID: "74e5397d-d0cb-4f02-890a-e59df5fdf7fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:07:42 crc kubenswrapper[4871]: I1007 23:07:42.556687 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74e5397d-d0cb-4f02-890a-e59df5fdf7fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:07:42 crc kubenswrapper[4871]: I1007 23:07:42.640191 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8kfc"] Oct 07 23:07:42 crc kubenswrapper[4871]: I1007 23:07:42.657109 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w8kfc"] Oct 07 23:07:42 crc kubenswrapper[4871]: I1007 23:07:42.998511 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" path="/var/lib/kubelet/pods/74e5397d-d0cb-4f02-890a-e59df5fdf7fa/volumes" Oct 07 23:08:05 crc kubenswrapper[4871]: I1007 23:08:05.512169 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:08:05 crc kubenswrapper[4871]: I1007 23:08:05.512932 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:08:35 crc kubenswrapper[4871]: I1007 23:08:35.512704 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:08:35 crc kubenswrapper[4871]: I1007 23:08:35.513668 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:09:05 crc kubenswrapper[4871]: I1007 23:09:05.512188 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:09:05 crc kubenswrapper[4871]: I1007 23:09:05.512939 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:09:05 crc kubenswrapper[4871]: I1007 23:09:05.513010 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:09:05 crc kubenswrapper[4871]: I1007 23:09:05.513962 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:09:05 crc kubenswrapper[4871]: I1007 23:09:05.514056 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" gracePeriod=600 Oct 07 23:09:05 crc kubenswrapper[4871]: E1007 23:09:05.643444 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:09:05 crc kubenswrapper[4871]: I1007 23:09:05.905370 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" exitCode=0 Oct 07 23:09:05 crc kubenswrapper[4871]: I1007 23:09:05.905437 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a"} Oct 07 23:09:05 crc kubenswrapper[4871]: I1007 23:09:05.905497 4871 scope.go:117] "RemoveContainer" containerID="4b53bd460ead0087e0d627e65163948ecfc09a4a5d94d569be0e1eeaab97ec3d" Oct 07 23:09:05 crc kubenswrapper[4871]: I1007 23:09:05.906218 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:09:05 crc kubenswrapper[4871]: E1007 23:09:05.906618 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:09:18 crc kubenswrapper[4871]: I1007 23:09:18.982075 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:09:18 crc kubenswrapper[4871]: E1007 23:09:18.983050 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:09:29 crc kubenswrapper[4871]: I1007 23:09:29.983643 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:09:29 crc kubenswrapper[4871]: E1007 23:09:29.984689 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:09:43 crc kubenswrapper[4871]: I1007 23:09:43.983316 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:09:43 crc kubenswrapper[4871]: E1007 23:09:43.984401 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:09:56 crc kubenswrapper[4871]: I1007 23:09:56.990638 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:09:56 crc kubenswrapper[4871]: E1007 23:09:56.991519 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:10:08 crc kubenswrapper[4871]: I1007 23:10:08.982177 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:10:08 crc kubenswrapper[4871]: E1007 23:10:08.983421 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:10:19 crc kubenswrapper[4871]: I1007 23:10:19.983079 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:10:19 crc kubenswrapper[4871]: E1007 23:10:19.984147 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:10:31 crc kubenswrapper[4871]: I1007 23:10:31.982228 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:10:31 crc kubenswrapper[4871]: E1007 23:10:31.983125 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:10:45 crc kubenswrapper[4871]: I1007 23:10:45.982447 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:10:45 crc kubenswrapper[4871]: E1007 23:10:45.983418 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:11:00 crc kubenswrapper[4871]: I1007 23:11:00.982096 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:11:00 crc kubenswrapper[4871]: E1007 23:11:00.982814 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:11:13 crc kubenswrapper[4871]: I1007 23:11:13.982037 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:11:13 crc kubenswrapper[4871]: E1007 23:11:13.982922 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:11:24 crc kubenswrapper[4871]: I1007 23:11:24.982542 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:11:24 crc kubenswrapper[4871]: E1007 23:11:24.983581 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.535309 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kztwf"] Oct 07 23:11:25 crc kubenswrapper[4871]: E1007 23:11:25.535739 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerName="extract-utilities" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.535759 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerName="extract-utilities" Oct 07 23:11:25 crc kubenswrapper[4871]: E1007 23:11:25.535788 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerName="registry-server" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.535829 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerName="registry-server" Oct 07 23:11:25 crc kubenswrapper[4871]: E1007 23:11:25.535876 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerName="extract-content" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.535888 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerName="extract-content" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.536182 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="74e5397d-d0cb-4f02-890a-e59df5fdf7fa" containerName="registry-server" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.537999 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.560209 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kztwf"] Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.572396 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-catalog-content\") pod \"certified-operators-kztwf\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.572520 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-utilities\") pod \"certified-operators-kztwf\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.572641 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgdvc\" (UniqueName: \"kubernetes.io/projected/3fc3975b-2e5e-43c8-9870-00e62b893bc8-kube-api-access-hgdvc\") pod \"certified-operators-kztwf\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.673992 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-utilities\") pod \"certified-operators-kztwf\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.674156 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgdvc\" (UniqueName: \"kubernetes.io/projected/3fc3975b-2e5e-43c8-9870-00e62b893bc8-kube-api-access-hgdvc\") pod \"certified-operators-kztwf\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.674228 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-catalog-content\") pod \"certified-operators-kztwf\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.677690 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-utilities\") pod \"certified-operators-kztwf\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.677779 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-catalog-content\") pod \"certified-operators-kztwf\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.697504 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgdvc\" (UniqueName: \"kubernetes.io/projected/3fc3975b-2e5e-43c8-9870-00e62b893bc8-kube-api-access-hgdvc\") pod \"certified-operators-kztwf\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:25 crc kubenswrapper[4871]: I1007 23:11:25.872782 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:26 crc kubenswrapper[4871]: I1007 23:11:26.340428 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kztwf"] Oct 07 23:11:27 crc kubenswrapper[4871]: I1007 23:11:27.175759 4871 generic.go:334] "Generic (PLEG): container finished" podID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerID="2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b" exitCode=0 Oct 07 23:11:27 crc kubenswrapper[4871]: I1007 23:11:27.175915 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztwf" event={"ID":"3fc3975b-2e5e-43c8-9870-00e62b893bc8","Type":"ContainerDied","Data":"2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b"} Oct 07 23:11:27 crc kubenswrapper[4871]: I1007 23:11:27.175973 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztwf" event={"ID":"3fc3975b-2e5e-43c8-9870-00e62b893bc8","Type":"ContainerStarted","Data":"9affea43f7ed517abf0ac1ceddbab1268f555e8460401f935ed7eb1a77424323"} Oct 07 23:11:29 crc kubenswrapper[4871]: I1007 23:11:29.197885 4871 generic.go:334] "Generic (PLEG): container finished" podID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerID="d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14" exitCode=0 Oct 07 23:11:29 crc kubenswrapper[4871]: I1007 23:11:29.197964 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztwf" event={"ID":"3fc3975b-2e5e-43c8-9870-00e62b893bc8","Type":"ContainerDied","Data":"d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14"} Oct 07 23:11:30 crc kubenswrapper[4871]: I1007 23:11:30.210260 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztwf" event={"ID":"3fc3975b-2e5e-43c8-9870-00e62b893bc8","Type":"ContainerStarted","Data":"579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f"} Oct 07 23:11:30 crc kubenswrapper[4871]: I1007 23:11:30.232464 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kztwf" podStartSLOduration=2.721834521 podStartE2EDuration="5.232447754s" podCreationTimestamp="2025-10-07 23:11:25 +0000 UTC" firstStartedPulling="2025-10-07 23:11:27.17939516 +0000 UTC m=+3760.982093273" lastFinishedPulling="2025-10-07 23:11:29.690008413 +0000 UTC m=+3763.492706506" observedRunningTime="2025-10-07 23:11:30.23114032 +0000 UTC m=+3764.033838393" watchObservedRunningTime="2025-10-07 23:11:30.232447754 +0000 UTC m=+3764.035145817" Oct 07 23:11:35 crc kubenswrapper[4871]: I1007 23:11:35.873936 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:35 crc kubenswrapper[4871]: I1007 23:11:35.875023 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:35 crc kubenswrapper[4871]: I1007 23:11:35.946144 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:36 crc kubenswrapper[4871]: I1007 23:11:36.331500 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:36 crc kubenswrapper[4871]: I1007 23:11:36.403460 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kztwf"] Oct 07 23:11:38 crc kubenswrapper[4871]: I1007 23:11:38.284757 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kztwf" podUID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerName="registry-server" containerID="cri-o://579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f" gracePeriod=2 Oct 07 23:11:38 crc kubenswrapper[4871]: I1007 23:11:38.787539 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:38 crc kubenswrapper[4871]: I1007 23:11:38.982314 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:11:38 crc kubenswrapper[4871]: E1007 23:11:38.982847 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:11:38 crc kubenswrapper[4871]: I1007 23:11:38.985478 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-utilities\") pod \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " Oct 07 23:11:38 crc kubenswrapper[4871]: I1007 23:11:38.985644 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgdvc\" (UniqueName: \"kubernetes.io/projected/3fc3975b-2e5e-43c8-9870-00e62b893bc8-kube-api-access-hgdvc\") pod \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " Oct 07 23:11:38 crc kubenswrapper[4871]: I1007 23:11:38.985763 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-catalog-content\") pod \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\" (UID: \"3fc3975b-2e5e-43c8-9870-00e62b893bc8\") " Oct 07 23:11:38 crc kubenswrapper[4871]: I1007 23:11:38.988013 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-utilities" (OuterVolumeSpecName: "utilities") pod "3fc3975b-2e5e-43c8-9870-00e62b893bc8" (UID: "3fc3975b-2e5e-43c8-9870-00e62b893bc8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:11:38 crc kubenswrapper[4871]: I1007 23:11:38.997967 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fc3975b-2e5e-43c8-9870-00e62b893bc8-kube-api-access-hgdvc" (OuterVolumeSpecName: "kube-api-access-hgdvc") pod "3fc3975b-2e5e-43c8-9870-00e62b893bc8" (UID: "3fc3975b-2e5e-43c8-9870-00e62b893bc8"). InnerVolumeSpecName "kube-api-access-hgdvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.087755 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.087835 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgdvc\" (UniqueName: \"kubernetes.io/projected/3fc3975b-2e5e-43c8-9870-00e62b893bc8-kube-api-access-hgdvc\") on node \"crc\" DevicePath \"\"" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.148550 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3fc3975b-2e5e-43c8-9870-00e62b893bc8" (UID: "3fc3975b-2e5e-43c8-9870-00e62b893bc8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.189151 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fc3975b-2e5e-43c8-9870-00e62b893bc8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.294254 4871 generic.go:334] "Generic (PLEG): container finished" podID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerID="579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f" exitCode=0 Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.294299 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztwf" event={"ID":"3fc3975b-2e5e-43c8-9870-00e62b893bc8","Type":"ContainerDied","Data":"579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f"} Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.294326 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztwf" event={"ID":"3fc3975b-2e5e-43c8-9870-00e62b893bc8","Type":"ContainerDied","Data":"9affea43f7ed517abf0ac1ceddbab1268f555e8460401f935ed7eb1a77424323"} Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.294341 4871 scope.go:117] "RemoveContainer" containerID="579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.296092 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kztwf" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.332158 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kztwf"] Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.334768 4871 scope.go:117] "RemoveContainer" containerID="d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.342333 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kztwf"] Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.367749 4871 scope.go:117] "RemoveContainer" containerID="2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.393110 4871 scope.go:117] "RemoveContainer" containerID="579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f" Oct 07 23:11:39 crc kubenswrapper[4871]: E1007 23:11:39.393620 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f\": container with ID starting with 579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f not found: ID does not exist" containerID="579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.393666 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f"} err="failed to get container status \"579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f\": rpc error: code = NotFound desc = could not find container \"579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f\": container with ID starting with 579e38a8d5778b046371953648c76dd590f799fdc6ef1cb99702989ffc662f8f not found: ID does not exist" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.393693 4871 scope.go:117] "RemoveContainer" containerID="d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14" Oct 07 23:11:39 crc kubenswrapper[4871]: E1007 23:11:39.394084 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14\": container with ID starting with d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14 not found: ID does not exist" containerID="d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.394158 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14"} err="failed to get container status \"d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14\": rpc error: code = NotFound desc = could not find container \"d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14\": container with ID starting with d8ebf31b3126fd04c658e95fe026175786917b2084d6625fa08633d74d71be14 not found: ID does not exist" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.394197 4871 scope.go:117] "RemoveContainer" containerID="2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b" Oct 07 23:11:39 crc kubenswrapper[4871]: E1007 23:11:39.394517 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b\": container with ID starting with 2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b not found: ID does not exist" containerID="2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b" Oct 07 23:11:39 crc kubenswrapper[4871]: I1007 23:11:39.394571 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b"} err="failed to get container status \"2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b\": rpc error: code = NotFound desc = could not find container \"2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b\": container with ID starting with 2e673efebc58dd5b87bdb91549927c25437862f399188252caa1539e27fc844b not found: ID does not exist" Oct 07 23:11:40 crc kubenswrapper[4871]: I1007 23:11:40.993145 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" path="/var/lib/kubelet/pods/3fc3975b-2e5e-43c8-9870-00e62b893bc8/volumes" Oct 07 23:11:49 crc kubenswrapper[4871]: I1007 23:11:49.982083 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:11:49 crc kubenswrapper[4871]: E1007 23:11:49.983057 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:12:02 crc kubenswrapper[4871]: I1007 23:12:02.982993 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:12:02 crc kubenswrapper[4871]: E1007 23:12:02.983926 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:12:13 crc kubenswrapper[4871]: I1007 23:12:13.982498 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:12:13 crc kubenswrapper[4871]: E1007 23:12:13.984475 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:12:28 crc kubenswrapper[4871]: I1007 23:12:28.982213 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:12:28 crc kubenswrapper[4871]: E1007 23:12:28.983465 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:12:42 crc kubenswrapper[4871]: I1007 23:12:42.982875 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:12:42 crc kubenswrapper[4871]: E1007 23:12:42.984178 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:12:55 crc kubenswrapper[4871]: I1007 23:12:55.982132 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:12:55 crc kubenswrapper[4871]: E1007 23:12:55.983030 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:13:08 crc kubenswrapper[4871]: I1007 23:13:08.983175 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:13:08 crc kubenswrapper[4871]: E1007 23:13:08.983926 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:13:23 crc kubenswrapper[4871]: I1007 23:13:23.983209 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:13:23 crc kubenswrapper[4871]: E1007 23:13:23.984398 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:13:38 crc kubenswrapper[4871]: I1007 23:13:38.983814 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:13:38 crc kubenswrapper[4871]: E1007 23:13:38.985043 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:13:49 crc kubenswrapper[4871]: I1007 23:13:49.983116 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:13:49 crc kubenswrapper[4871]: E1007 23:13:49.984180 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:14:01 crc kubenswrapper[4871]: I1007 23:14:01.982141 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:14:01 crc kubenswrapper[4871]: E1007 23:14:01.982757 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:14:16 crc kubenswrapper[4871]: I1007 23:14:16.986004 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:14:17 crc kubenswrapper[4871]: I1007 23:14:17.760974 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"ba8382184c0497f4ecbce3a39a20b51e984bb7a756965744ae29bcb76bcad67f"} Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.155531 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9"] Oct 07 23:15:00 crc kubenswrapper[4871]: E1007 23:15:00.156274 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerName="extract-content" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.156287 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerName="extract-content" Oct 07 23:15:00 crc kubenswrapper[4871]: E1007 23:15:00.156295 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerName="registry-server" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.156302 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerName="registry-server" Oct 07 23:15:00 crc kubenswrapper[4871]: E1007 23:15:00.156315 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerName="extract-utilities" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.156321 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerName="extract-utilities" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.156439 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc3975b-2e5e-43c8-9870-00e62b893bc8" containerName="registry-server" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.158182 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.160164 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.162328 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.174331 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9"] Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.344233 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-config-volume\") pod \"collect-profiles-29331315-76nm9\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.344297 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-secret-volume\") pod \"collect-profiles-29331315-76nm9\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.344368 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq7bp\" (UniqueName: \"kubernetes.io/projected/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-kube-api-access-gq7bp\") pod \"collect-profiles-29331315-76nm9\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.446169 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-config-volume\") pod \"collect-profiles-29331315-76nm9\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.446232 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-secret-volume\") pod \"collect-profiles-29331315-76nm9\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.446303 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq7bp\" (UniqueName: \"kubernetes.io/projected/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-kube-api-access-gq7bp\") pod \"collect-profiles-29331315-76nm9\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.448754 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-config-volume\") pod \"collect-profiles-29331315-76nm9\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.456785 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-secret-volume\") pod \"collect-profiles-29331315-76nm9\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.468478 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq7bp\" (UniqueName: \"kubernetes.io/projected/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-kube-api-access-gq7bp\") pod \"collect-profiles-29331315-76nm9\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:00 crc kubenswrapper[4871]: I1007 23:15:00.502232 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:01 crc kubenswrapper[4871]: I1007 23:15:01.045859 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9"] Oct 07 23:15:01 crc kubenswrapper[4871]: I1007 23:15:01.171084 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" event={"ID":"2d395bf3-85d2-4f95-af96-d62c4bc13ab0","Type":"ContainerStarted","Data":"5946c9b3b2d2b1e85cb9e8a4d685e0895273174cb87cd4b182f258427071d381"} Oct 07 23:15:02 crc kubenswrapper[4871]: I1007 23:15:02.182914 4871 generic.go:334] "Generic (PLEG): container finished" podID="2d395bf3-85d2-4f95-af96-d62c4bc13ab0" containerID="f6e7c38f0f0b0e1e2e7b44d963307c35a3e49bed289b3b56d39cd1c6213331b0" exitCode=0 Oct 07 23:15:02 crc kubenswrapper[4871]: I1007 23:15:02.182997 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" event={"ID":"2d395bf3-85d2-4f95-af96-d62c4bc13ab0","Type":"ContainerDied","Data":"f6e7c38f0f0b0e1e2e7b44d963307c35a3e49bed289b3b56d39cd1c6213331b0"} Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.588023 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.594636 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-config-volume\") pod \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.594717 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq7bp\" (UniqueName: \"kubernetes.io/projected/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-kube-api-access-gq7bp\") pod \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.594769 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-secret-volume\") pod \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\" (UID: \"2d395bf3-85d2-4f95-af96-d62c4bc13ab0\") " Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.596040 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-config-volume" (OuterVolumeSpecName: "config-volume") pod "2d395bf3-85d2-4f95-af96-d62c4bc13ab0" (UID: "2d395bf3-85d2-4f95-af96-d62c4bc13ab0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.604104 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-kube-api-access-gq7bp" (OuterVolumeSpecName: "kube-api-access-gq7bp") pod "2d395bf3-85d2-4f95-af96-d62c4bc13ab0" (UID: "2d395bf3-85d2-4f95-af96-d62c4bc13ab0"). InnerVolumeSpecName "kube-api-access-gq7bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.604121 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2d395bf3-85d2-4f95-af96-d62c4bc13ab0" (UID: "2d395bf3-85d2-4f95-af96-d62c4bc13ab0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.696840 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.696889 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq7bp\" (UniqueName: \"kubernetes.io/projected/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-kube-api-access-gq7bp\") on node \"crc\" DevicePath \"\"" Oct 07 23:15:03 crc kubenswrapper[4871]: I1007 23:15:03.696908 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d395bf3-85d2-4f95-af96-d62c4bc13ab0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 23:15:04 crc kubenswrapper[4871]: I1007 23:15:04.202196 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" event={"ID":"2d395bf3-85d2-4f95-af96-d62c4bc13ab0","Type":"ContainerDied","Data":"5946c9b3b2d2b1e85cb9e8a4d685e0895273174cb87cd4b182f258427071d381"} Oct 07 23:15:04 crc kubenswrapper[4871]: I1007 23:15:04.202528 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5946c9b3b2d2b1e85cb9e8a4d685e0895273174cb87cd4b182f258427071d381" Oct 07 23:15:04 crc kubenswrapper[4871]: I1007 23:15:04.202273 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9" Oct 07 23:15:04 crc kubenswrapper[4871]: I1007 23:15:04.690914 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc"] Oct 07 23:15:04 crc kubenswrapper[4871]: I1007 23:15:04.695710 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331270-d9xzc"] Oct 07 23:15:04 crc kubenswrapper[4871]: I1007 23:15:04.996553 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21aa285c-6d29-4ef3-b846-bc1f02fc5416" path="/var/lib/kubelet/pods/21aa285c-6d29-4ef3-b846-bc1f02fc5416/volumes" Oct 07 23:15:55 crc kubenswrapper[4871]: I1007 23:15:55.368967 4871 scope.go:117] "RemoveContainer" containerID="d91dafd4d911f590589061e8ce31ae1a7587033551714a110604605d111388b2" Oct 07 23:16:35 crc kubenswrapper[4871]: I1007 23:16:35.513031 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:16:35 crc kubenswrapper[4871]: I1007 23:16:35.513688 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:17:05 crc kubenswrapper[4871]: I1007 23:17:05.512459 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:17:05 crc kubenswrapper[4871]: I1007 23:17:05.513158 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:17:35 crc kubenswrapper[4871]: I1007 23:17:35.512301 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:17:35 crc kubenswrapper[4871]: I1007 23:17:35.513000 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:17:35 crc kubenswrapper[4871]: I1007 23:17:35.513062 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:17:35 crc kubenswrapper[4871]: I1007 23:17:35.514043 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba8382184c0497f4ecbce3a39a20b51e984bb7a756965744ae29bcb76bcad67f"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:17:35 crc kubenswrapper[4871]: I1007 23:17:35.514146 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://ba8382184c0497f4ecbce3a39a20b51e984bb7a756965744ae29bcb76bcad67f" gracePeriod=600 Oct 07 23:17:35 crc kubenswrapper[4871]: I1007 23:17:35.691963 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="ba8382184c0497f4ecbce3a39a20b51e984bb7a756965744ae29bcb76bcad67f" exitCode=0 Oct 07 23:17:35 crc kubenswrapper[4871]: I1007 23:17:35.692036 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"ba8382184c0497f4ecbce3a39a20b51e984bb7a756965744ae29bcb76bcad67f"} Oct 07 23:17:35 crc kubenswrapper[4871]: I1007 23:17:35.692111 4871 scope.go:117] "RemoveContainer" containerID="7826b0b54ed9e08a4cc31a8a99ab197bf9c311d95bb71f92e0bcfcf50e90680a" Oct 07 23:17:36 crc kubenswrapper[4871]: I1007 23:17:36.706458 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e"} Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.437056 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nwzl7"] Oct 07 23:17:53 crc kubenswrapper[4871]: E1007 23:17:53.438195 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d395bf3-85d2-4f95-af96-d62c4bc13ab0" containerName="collect-profiles" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.438217 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d395bf3-85d2-4f95-af96-d62c4bc13ab0" containerName="collect-profiles" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.438508 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d395bf3-85d2-4f95-af96-d62c4bc13ab0" containerName="collect-profiles" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.440345 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.458057 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwzl7"] Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.520645 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-catalog-content\") pod \"redhat-marketplace-nwzl7\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.520706 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-utilities\") pod \"redhat-marketplace-nwzl7\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.520815 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4w7h\" (UniqueName: \"kubernetes.io/projected/9d75f845-d44a-4d17-87df-6468fe5ccdec-kube-api-access-c4w7h\") pod \"redhat-marketplace-nwzl7\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.622015 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4w7h\" (UniqueName: \"kubernetes.io/projected/9d75f845-d44a-4d17-87df-6468fe5ccdec-kube-api-access-c4w7h\") pod \"redhat-marketplace-nwzl7\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.622099 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-catalog-content\") pod \"redhat-marketplace-nwzl7\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.622137 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-utilities\") pod \"redhat-marketplace-nwzl7\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.622704 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-utilities\") pod \"redhat-marketplace-nwzl7\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.622973 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-catalog-content\") pod \"redhat-marketplace-nwzl7\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:53 crc kubenswrapper[4871]: I1007 23:17:53.880953 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4w7h\" (UniqueName: \"kubernetes.io/projected/9d75f845-d44a-4d17-87df-6468fe5ccdec-kube-api-access-c4w7h\") pod \"redhat-marketplace-nwzl7\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:54 crc kubenswrapper[4871]: I1007 23:17:54.070442 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:17:54 crc kubenswrapper[4871]: I1007 23:17:54.541266 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwzl7"] Oct 07 23:17:54 crc kubenswrapper[4871]: I1007 23:17:54.877872 4871 generic.go:334] "Generic (PLEG): container finished" podID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerID="2e410a288b0644e77d7d50d658c333ceffd37e406e932bcbcbe9a3c3c8c7c744" exitCode=0 Oct 07 23:17:54 crc kubenswrapper[4871]: I1007 23:17:54.877939 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwzl7" event={"ID":"9d75f845-d44a-4d17-87df-6468fe5ccdec","Type":"ContainerDied","Data":"2e410a288b0644e77d7d50d658c333ceffd37e406e932bcbcbe9a3c3c8c7c744"} Oct 07 23:17:54 crc kubenswrapper[4871]: I1007 23:17:54.877976 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwzl7" event={"ID":"9d75f845-d44a-4d17-87df-6468fe5ccdec","Type":"ContainerStarted","Data":"fe6c1b6278321ceaa1014751418d0c9a3edf159c90ae355ef89931c2eb4566df"} Oct 07 23:17:54 crc kubenswrapper[4871]: I1007 23:17:54.880306 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 23:17:56 crc kubenswrapper[4871]: I1007 23:17:56.905430 4871 generic.go:334] "Generic (PLEG): container finished" podID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerID="44944b2b560be5ba8cda32597e753207f059bb65f38c89d93361cb4bc02af202" exitCode=0 Oct 07 23:17:56 crc kubenswrapper[4871]: I1007 23:17:56.905521 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwzl7" event={"ID":"9d75f845-d44a-4d17-87df-6468fe5ccdec","Type":"ContainerDied","Data":"44944b2b560be5ba8cda32597e753207f059bb65f38c89d93361cb4bc02af202"} Oct 07 23:17:57 crc kubenswrapper[4871]: I1007 23:17:57.917600 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwzl7" event={"ID":"9d75f845-d44a-4d17-87df-6468fe5ccdec","Type":"ContainerStarted","Data":"23546bfd432fd6f5e99b8618904f405d15eba474ea295730b94ee11ad7d99a04"} Oct 07 23:17:57 crc kubenswrapper[4871]: I1007 23:17:57.939919 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nwzl7" podStartSLOduration=2.494374115 podStartE2EDuration="4.939905397s" podCreationTimestamp="2025-10-07 23:17:53 +0000 UTC" firstStartedPulling="2025-10-07 23:17:54.880074699 +0000 UTC m=+4148.682772772" lastFinishedPulling="2025-10-07 23:17:57.325605951 +0000 UTC m=+4151.128304054" observedRunningTime="2025-10-07 23:17:57.938576482 +0000 UTC m=+4151.741274565" watchObservedRunningTime="2025-10-07 23:17:57.939905397 +0000 UTC m=+4151.742603470" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.608262 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wj4l5"] Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.611013 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.632609 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wj4l5"] Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.721433 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-utilities\") pod \"redhat-operators-wj4l5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.721630 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-catalog-content\") pod \"redhat-operators-wj4l5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.721682 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsvx4\" (UniqueName: \"kubernetes.io/projected/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-kube-api-access-dsvx4\") pod \"redhat-operators-wj4l5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.823623 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-catalog-content\") pod \"redhat-operators-wj4l5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.823707 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsvx4\" (UniqueName: \"kubernetes.io/projected/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-kube-api-access-dsvx4\") pod \"redhat-operators-wj4l5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.823878 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-utilities\") pod \"redhat-operators-wj4l5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.824599 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-catalog-content\") pod \"redhat-operators-wj4l5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.824733 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-utilities\") pod \"redhat-operators-wj4l5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.855148 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsvx4\" (UniqueName: \"kubernetes.io/projected/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-kube-api-access-dsvx4\") pod \"redhat-operators-wj4l5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:17:59 crc kubenswrapper[4871]: I1007 23:17:59.950404 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:18:00 crc kubenswrapper[4871]: I1007 23:18:00.205464 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wj4l5"] Oct 07 23:18:00 crc kubenswrapper[4871]: W1007 23:18:00.211997 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0225c14_cffa_48ce_9f79_8f9e5ead0fd5.slice/crio-c7265dbfba2e7ebf24651cd2309fa24aefe225cc3d67e4b70a52877382369b2f WatchSource:0}: Error finding container c7265dbfba2e7ebf24651cd2309fa24aefe225cc3d67e4b70a52877382369b2f: Status 404 returned error can't find the container with id c7265dbfba2e7ebf24651cd2309fa24aefe225cc3d67e4b70a52877382369b2f Oct 07 23:18:00 crc kubenswrapper[4871]: I1007 23:18:00.952473 4871 generic.go:334] "Generic (PLEG): container finished" podID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerID="62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2" exitCode=0 Oct 07 23:18:00 crc kubenswrapper[4871]: I1007 23:18:00.952929 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wj4l5" event={"ID":"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5","Type":"ContainerDied","Data":"62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2"} Oct 07 23:18:00 crc kubenswrapper[4871]: I1007 23:18:00.952976 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wj4l5" event={"ID":"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5","Type":"ContainerStarted","Data":"c7265dbfba2e7ebf24651cd2309fa24aefe225cc3d67e4b70a52877382369b2f"} Oct 07 23:18:02 crc kubenswrapper[4871]: I1007 23:18:02.974668 4871 generic.go:334] "Generic (PLEG): container finished" podID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerID="9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708" exitCode=0 Oct 07 23:18:02 crc kubenswrapper[4871]: I1007 23:18:02.974846 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wj4l5" event={"ID":"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5","Type":"ContainerDied","Data":"9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708"} Oct 07 23:18:03 crc kubenswrapper[4871]: I1007 23:18:03.989884 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wj4l5" event={"ID":"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5","Type":"ContainerStarted","Data":"77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719"} Oct 07 23:18:04 crc kubenswrapper[4871]: I1007 23:18:04.071626 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:18:04 crc kubenswrapper[4871]: I1007 23:18:04.071703 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:18:04 crc kubenswrapper[4871]: I1007 23:18:04.135423 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:18:04 crc kubenswrapper[4871]: I1007 23:18:04.161657 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wj4l5" podStartSLOduration=2.644291797 podStartE2EDuration="5.161630954s" podCreationTimestamp="2025-10-07 23:17:59 +0000 UTC" firstStartedPulling="2025-10-07 23:18:00.955924748 +0000 UTC m=+4154.758622861" lastFinishedPulling="2025-10-07 23:18:03.473263915 +0000 UTC m=+4157.275962018" observedRunningTime="2025-10-07 23:18:04.016018522 +0000 UTC m=+4157.818716636" watchObservedRunningTime="2025-10-07 23:18:04.161630954 +0000 UTC m=+4157.964329057" Oct 07 23:18:05 crc kubenswrapper[4871]: I1007 23:18:05.070672 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:18:06 crc kubenswrapper[4871]: I1007 23:18:06.002696 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwzl7"] Oct 07 23:18:07 crc kubenswrapper[4871]: I1007 23:18:07.017334 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nwzl7" podUID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerName="registry-server" containerID="cri-o://23546bfd432fd6f5e99b8618904f405d15eba474ea295730b94ee11ad7d99a04" gracePeriod=2 Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.027139 4871 generic.go:334] "Generic (PLEG): container finished" podID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerID="23546bfd432fd6f5e99b8618904f405d15eba474ea295730b94ee11ad7d99a04" exitCode=0 Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.027212 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwzl7" event={"ID":"9d75f845-d44a-4d17-87df-6468fe5ccdec","Type":"ContainerDied","Data":"23546bfd432fd6f5e99b8618904f405d15eba474ea295730b94ee11ad7d99a04"} Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.027436 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwzl7" event={"ID":"9d75f845-d44a-4d17-87df-6468fe5ccdec","Type":"ContainerDied","Data":"fe6c1b6278321ceaa1014751418d0c9a3edf159c90ae355ef89931c2eb4566df"} Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.027448 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe6c1b6278321ceaa1014751418d0c9a3edf159c90ae355ef89931c2eb4566df" Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.064470 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.173498 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4w7h\" (UniqueName: \"kubernetes.io/projected/9d75f845-d44a-4d17-87df-6468fe5ccdec-kube-api-access-c4w7h\") pod \"9d75f845-d44a-4d17-87df-6468fe5ccdec\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.173569 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-utilities\") pod \"9d75f845-d44a-4d17-87df-6468fe5ccdec\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.173674 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-catalog-content\") pod \"9d75f845-d44a-4d17-87df-6468fe5ccdec\" (UID: \"9d75f845-d44a-4d17-87df-6468fe5ccdec\") " Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.174739 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-utilities" (OuterVolumeSpecName: "utilities") pod "9d75f845-d44a-4d17-87df-6468fe5ccdec" (UID: "9d75f845-d44a-4d17-87df-6468fe5ccdec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.179246 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d75f845-d44a-4d17-87df-6468fe5ccdec-kube-api-access-c4w7h" (OuterVolumeSpecName: "kube-api-access-c4w7h") pod "9d75f845-d44a-4d17-87df-6468fe5ccdec" (UID: "9d75f845-d44a-4d17-87df-6468fe5ccdec"). InnerVolumeSpecName "kube-api-access-c4w7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.196750 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d75f845-d44a-4d17-87df-6468fe5ccdec" (UID: "9d75f845-d44a-4d17-87df-6468fe5ccdec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.274854 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4w7h\" (UniqueName: \"kubernetes.io/projected/9d75f845-d44a-4d17-87df-6468fe5ccdec-kube-api-access-c4w7h\") on node \"crc\" DevicePath \"\"" Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.274890 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:18:08 crc kubenswrapper[4871]: I1007 23:18:08.274899 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d75f845-d44a-4d17-87df-6468fe5ccdec-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:18:09 crc kubenswrapper[4871]: I1007 23:18:09.032817 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwzl7" Oct 07 23:18:09 crc kubenswrapper[4871]: I1007 23:18:09.063719 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwzl7"] Oct 07 23:18:09 crc kubenswrapper[4871]: I1007 23:18:09.073083 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwzl7"] Oct 07 23:18:09 crc kubenswrapper[4871]: I1007 23:18:09.950990 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:18:09 crc kubenswrapper[4871]: I1007 23:18:09.951072 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:18:10 crc kubenswrapper[4871]: I1007 23:18:10.005972 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:18:10 crc kubenswrapper[4871]: I1007 23:18:10.088739 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:18:10 crc kubenswrapper[4871]: I1007 23:18:10.998046 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d75f845-d44a-4d17-87df-6468fe5ccdec" path="/var/lib/kubelet/pods/9d75f845-d44a-4d17-87df-6468fe5ccdec/volumes" Oct 07 23:18:11 crc kubenswrapper[4871]: I1007 23:18:11.399282 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wj4l5"] Oct 07 23:18:12 crc kubenswrapper[4871]: I1007 23:18:12.055237 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wj4l5" podUID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerName="registry-server" containerID="cri-o://77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719" gracePeriod=2 Oct 07 23:18:12 crc kubenswrapper[4871]: I1007 23:18:12.524487 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:18:12 crc kubenswrapper[4871]: I1007 23:18:12.643130 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-utilities\") pod \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " Oct 07 23:18:12 crc kubenswrapper[4871]: I1007 23:18:12.643191 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-catalog-content\") pod \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " Oct 07 23:18:12 crc kubenswrapper[4871]: I1007 23:18:12.643259 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsvx4\" (UniqueName: \"kubernetes.io/projected/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-kube-api-access-dsvx4\") pod \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\" (UID: \"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5\") " Oct 07 23:18:12 crc kubenswrapper[4871]: I1007 23:18:12.643993 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-utilities" (OuterVolumeSpecName: "utilities") pod "b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" (UID: "b0225c14-cffa-48ce-9f79-8f9e5ead0fd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:18:12 crc kubenswrapper[4871]: I1007 23:18:12.659190 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-kube-api-access-dsvx4" (OuterVolumeSpecName: "kube-api-access-dsvx4") pod "b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" (UID: "b0225c14-cffa-48ce-9f79-8f9e5ead0fd5"). InnerVolumeSpecName "kube-api-access-dsvx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:18:12 crc kubenswrapper[4871]: I1007 23:18:12.745990 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:18:12 crc kubenswrapper[4871]: I1007 23:18:12.746053 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsvx4\" (UniqueName: \"kubernetes.io/projected/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-kube-api-access-dsvx4\") on node \"crc\" DevicePath \"\"" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.069173 4871 generic.go:334] "Generic (PLEG): container finished" podID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerID="77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719" exitCode=0 Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.069235 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wj4l5" event={"ID":"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5","Type":"ContainerDied","Data":"77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719"} Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.069276 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wj4l5" event={"ID":"b0225c14-cffa-48ce-9f79-8f9e5ead0fd5","Type":"ContainerDied","Data":"c7265dbfba2e7ebf24651cd2309fa24aefe225cc3d67e4b70a52877382369b2f"} Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.069306 4871 scope.go:117] "RemoveContainer" containerID="77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.069517 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wj4l5" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.099706 4871 scope.go:117] "RemoveContainer" containerID="9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.124701 4871 scope.go:117] "RemoveContainer" containerID="62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.165644 4871 scope.go:117] "RemoveContainer" containerID="77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719" Oct 07 23:18:13 crc kubenswrapper[4871]: E1007 23:18:13.166243 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719\": container with ID starting with 77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719 not found: ID does not exist" containerID="77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.166306 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719"} err="failed to get container status \"77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719\": rpc error: code = NotFound desc = could not find container \"77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719\": container with ID starting with 77de0ec4505ed377f1f4aa56ad88ea0374b655381b25364f81fa8a41606de719 not found: ID does not exist" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.166339 4871 scope.go:117] "RemoveContainer" containerID="9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708" Oct 07 23:18:13 crc kubenswrapper[4871]: E1007 23:18:13.166897 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708\": container with ID starting with 9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708 not found: ID does not exist" containerID="9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.166950 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708"} err="failed to get container status \"9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708\": rpc error: code = NotFound desc = could not find container \"9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708\": container with ID starting with 9c4e9c3ca71270ec691bba12b43512540fda0252e80ee6e8d73d9c3807ed9708 not found: ID does not exist" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.166989 4871 scope.go:117] "RemoveContainer" containerID="62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2" Oct 07 23:18:13 crc kubenswrapper[4871]: E1007 23:18:13.167376 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2\": container with ID starting with 62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2 not found: ID does not exist" containerID="62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.167411 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2"} err="failed to get container status \"62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2\": rpc error: code = NotFound desc = could not find container \"62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2\": container with ID starting with 62df5a0447b98da99ed0503e4ccad411c44eaec565ae9a9db8ceb4e567b4dbc2 not found: ID does not exist" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.800450 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" (UID: "b0225c14-cffa-48ce-9f79-8f9e5ead0fd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:18:13 crc kubenswrapper[4871]: I1007 23:18:13.863526 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:18:14 crc kubenswrapper[4871]: I1007 23:18:14.019370 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wj4l5"] Oct 07 23:18:14 crc kubenswrapper[4871]: I1007 23:18:14.029993 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wj4l5"] Oct 07 23:18:15 crc kubenswrapper[4871]: I1007 23:18:15.002825 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" path="/var/lib/kubelet/pods/b0225c14-cffa-48ce-9f79-8f9e5ead0fd5/volumes" Oct 07 23:19:35 crc kubenswrapper[4871]: I1007 23:19:35.512056 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:19:35 crc kubenswrapper[4871]: I1007 23:19:35.512690 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:20:05 crc kubenswrapper[4871]: I1007 23:20:05.512742 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:20:05 crc kubenswrapper[4871]: I1007 23:20:05.513426 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:20:35 crc kubenswrapper[4871]: I1007 23:20:35.516015 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:20:35 crc kubenswrapper[4871]: I1007 23:20:35.517847 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:20:35 crc kubenswrapper[4871]: I1007 23:20:35.517979 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:20:35 crc kubenswrapper[4871]: I1007 23:20:35.518611 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:20:35 crc kubenswrapper[4871]: I1007 23:20:35.518731 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" gracePeriod=600 Oct 07 23:20:35 crc kubenswrapper[4871]: E1007 23:20:35.640176 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:20:36 crc kubenswrapper[4871]: I1007 23:20:36.462621 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" exitCode=0 Oct 07 23:20:36 crc kubenswrapper[4871]: I1007 23:20:36.462711 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e"} Oct 07 23:20:36 crc kubenswrapper[4871]: I1007 23:20:36.462852 4871 scope.go:117] "RemoveContainer" containerID="ba8382184c0497f4ecbce3a39a20b51e984bb7a756965744ae29bcb76bcad67f" Oct 07 23:20:36 crc kubenswrapper[4871]: I1007 23:20:36.463690 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:20:36 crc kubenswrapper[4871]: E1007 23:20:36.464238 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:20:49 crc kubenswrapper[4871]: I1007 23:20:49.982326 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:20:49 crc kubenswrapper[4871]: E1007 23:20:49.984998 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:21:04 crc kubenswrapper[4871]: I1007 23:21:04.982757 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:21:04 crc kubenswrapper[4871]: E1007 23:21:04.983834 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:21:19 crc kubenswrapper[4871]: I1007 23:21:19.982981 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:21:19 crc kubenswrapper[4871]: E1007 23:21:19.984008 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:21:33 crc kubenswrapper[4871]: I1007 23:21:33.982131 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:21:33 crc kubenswrapper[4871]: E1007 23:21:33.983066 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:21:44 crc kubenswrapper[4871]: I1007 23:21:44.982901 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:21:44 crc kubenswrapper[4871]: E1007 23:21:44.984219 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.486601 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l6c57"] Oct 07 23:21:49 crc kubenswrapper[4871]: E1007 23:21:49.490991 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerName="extract-content" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.491056 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerName="extract-content" Oct 07 23:21:49 crc kubenswrapper[4871]: E1007 23:21:49.491121 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerName="extract-utilities" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.491144 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerName="extract-utilities" Oct 07 23:21:49 crc kubenswrapper[4871]: E1007 23:21:49.491200 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerName="registry-server" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.491219 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerName="registry-server" Oct 07 23:21:49 crc kubenswrapper[4871]: E1007 23:21:49.491252 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerName="extract-utilities" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.491267 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerName="extract-utilities" Oct 07 23:21:49 crc kubenswrapper[4871]: E1007 23:21:49.491303 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerName="registry-server" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.491318 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerName="registry-server" Oct 07 23:21:49 crc kubenswrapper[4871]: E1007 23:21:49.491355 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerName="extract-content" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.491372 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerName="extract-content" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.491901 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d75f845-d44a-4d17-87df-6468fe5ccdec" containerName="registry-server" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.491955 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0225c14-cffa-48ce-9f79-8f9e5ead0fd5" containerName="registry-server" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.494434 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.513761 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l6c57"] Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.560098 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lswcl\" (UniqueName: \"kubernetes.io/projected/93a98e09-9cdd-4616-be66-1c362992fb02-kube-api-access-lswcl\") pod \"certified-operators-l6c57\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.560429 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-utilities\") pod \"certified-operators-l6c57\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.560681 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-catalog-content\") pod \"certified-operators-l6c57\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.661974 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-utilities\") pod \"certified-operators-l6c57\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.662073 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-catalog-content\") pod \"certified-operators-l6c57\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.662151 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lswcl\" (UniqueName: \"kubernetes.io/projected/93a98e09-9cdd-4616-be66-1c362992fb02-kube-api-access-lswcl\") pod \"certified-operators-l6c57\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.662516 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-catalog-content\") pod \"certified-operators-l6c57\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.662789 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-utilities\") pod \"certified-operators-l6c57\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.686357 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lswcl\" (UniqueName: \"kubernetes.io/projected/93a98e09-9cdd-4616-be66-1c362992fb02-kube-api-access-lswcl\") pod \"certified-operators-l6c57\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:49 crc kubenswrapper[4871]: I1007 23:21:49.837184 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:50 crc kubenswrapper[4871]: I1007 23:21:50.359436 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l6c57"] Oct 07 23:21:51 crc kubenswrapper[4871]: I1007 23:21:51.235212 4871 generic.go:334] "Generic (PLEG): container finished" podID="93a98e09-9cdd-4616-be66-1c362992fb02" containerID="91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5" exitCode=0 Oct 07 23:21:51 crc kubenswrapper[4871]: I1007 23:21:51.235366 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6c57" event={"ID":"93a98e09-9cdd-4616-be66-1c362992fb02","Type":"ContainerDied","Data":"91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5"} Oct 07 23:21:51 crc kubenswrapper[4871]: I1007 23:21:51.235611 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6c57" event={"ID":"93a98e09-9cdd-4616-be66-1c362992fb02","Type":"ContainerStarted","Data":"dc64d3cc0574eca46e69096933af147ab32d96ac431ac6ea6aea962d9ab8568a"} Oct 07 23:21:52 crc kubenswrapper[4871]: I1007 23:21:52.244724 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6c57" event={"ID":"93a98e09-9cdd-4616-be66-1c362992fb02","Type":"ContainerStarted","Data":"707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d"} Oct 07 23:21:53 crc kubenswrapper[4871]: I1007 23:21:53.257389 4871 generic.go:334] "Generic (PLEG): container finished" podID="93a98e09-9cdd-4616-be66-1c362992fb02" containerID="707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d" exitCode=0 Oct 07 23:21:53 crc kubenswrapper[4871]: I1007 23:21:53.257458 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6c57" event={"ID":"93a98e09-9cdd-4616-be66-1c362992fb02","Type":"ContainerDied","Data":"707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d"} Oct 07 23:21:54 crc kubenswrapper[4871]: I1007 23:21:54.269040 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6c57" event={"ID":"93a98e09-9cdd-4616-be66-1c362992fb02","Type":"ContainerStarted","Data":"d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58"} Oct 07 23:21:54 crc kubenswrapper[4871]: I1007 23:21:54.297809 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l6c57" podStartSLOduration=2.8887264569999997 podStartE2EDuration="5.297774607s" podCreationTimestamp="2025-10-07 23:21:49 +0000 UTC" firstStartedPulling="2025-10-07 23:21:51.23761504 +0000 UTC m=+4385.040313163" lastFinishedPulling="2025-10-07 23:21:53.64666321 +0000 UTC m=+4387.449361313" observedRunningTime="2025-10-07 23:21:54.294773477 +0000 UTC m=+4388.097471550" watchObservedRunningTime="2025-10-07 23:21:54.297774607 +0000 UTC m=+4388.100472680" Oct 07 23:21:55 crc kubenswrapper[4871]: I1007 23:21:55.983171 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:21:55 crc kubenswrapper[4871]: E1007 23:21:55.983983 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:21:59 crc kubenswrapper[4871]: I1007 23:21:59.837359 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:59 crc kubenswrapper[4871]: I1007 23:21:59.838145 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:21:59 crc kubenswrapper[4871]: I1007 23:21:59.915103 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:22:00 crc kubenswrapper[4871]: I1007 23:22:00.420899 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:22:00 crc kubenswrapper[4871]: I1007 23:22:00.490566 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l6c57"] Oct 07 23:22:02 crc kubenswrapper[4871]: I1007 23:22:02.357318 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l6c57" podUID="93a98e09-9cdd-4616-be66-1c362992fb02" containerName="registry-server" containerID="cri-o://d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58" gracePeriod=2 Oct 07 23:22:02 crc kubenswrapper[4871]: I1007 23:22:02.825407 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:22:02 crc kubenswrapper[4871]: I1007 23:22:02.982962 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lswcl\" (UniqueName: \"kubernetes.io/projected/93a98e09-9cdd-4616-be66-1c362992fb02-kube-api-access-lswcl\") pod \"93a98e09-9cdd-4616-be66-1c362992fb02\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " Oct 07 23:22:02 crc kubenswrapper[4871]: I1007 23:22:02.983177 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-utilities\") pod \"93a98e09-9cdd-4616-be66-1c362992fb02\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " Oct 07 23:22:02 crc kubenswrapper[4871]: I1007 23:22:02.983235 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-catalog-content\") pod \"93a98e09-9cdd-4616-be66-1c362992fb02\" (UID: \"93a98e09-9cdd-4616-be66-1c362992fb02\") " Oct 07 23:22:02 crc kubenswrapper[4871]: I1007 23:22:02.984852 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-utilities" (OuterVolumeSpecName: "utilities") pod "93a98e09-9cdd-4616-be66-1c362992fb02" (UID: "93a98e09-9cdd-4616-be66-1c362992fb02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:22:02 crc kubenswrapper[4871]: I1007 23:22:02.994099 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a98e09-9cdd-4616-be66-1c362992fb02-kube-api-access-lswcl" (OuterVolumeSpecName: "kube-api-access-lswcl") pod "93a98e09-9cdd-4616-be66-1c362992fb02" (UID: "93a98e09-9cdd-4616-be66-1c362992fb02"). InnerVolumeSpecName "kube-api-access-lswcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.063757 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93a98e09-9cdd-4616-be66-1c362992fb02" (UID: "93a98e09-9cdd-4616-be66-1c362992fb02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.085181 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.085225 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93a98e09-9cdd-4616-be66-1c362992fb02-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.085239 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lswcl\" (UniqueName: \"kubernetes.io/projected/93a98e09-9cdd-4616-be66-1c362992fb02-kube-api-access-lswcl\") on node \"crc\" DevicePath \"\"" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.369742 4871 generic.go:334] "Generic (PLEG): container finished" podID="93a98e09-9cdd-4616-be66-1c362992fb02" containerID="d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58" exitCode=0 Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.369846 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6c57" event={"ID":"93a98e09-9cdd-4616-be66-1c362992fb02","Type":"ContainerDied","Data":"d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58"} Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.369896 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l6c57" event={"ID":"93a98e09-9cdd-4616-be66-1c362992fb02","Type":"ContainerDied","Data":"dc64d3cc0574eca46e69096933af147ab32d96ac431ac6ea6aea962d9ab8568a"} Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.369921 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l6c57" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.369933 4871 scope.go:117] "RemoveContainer" containerID="d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.398941 4871 scope.go:117] "RemoveContainer" containerID="707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.419000 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l6c57"] Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.435546 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l6c57"] Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.446259 4871 scope.go:117] "RemoveContainer" containerID="91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.482411 4871 scope.go:117] "RemoveContainer" containerID="d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58" Oct 07 23:22:03 crc kubenswrapper[4871]: E1007 23:22:03.483135 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58\": container with ID starting with d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58 not found: ID does not exist" containerID="d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.483190 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58"} err="failed to get container status \"d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58\": rpc error: code = NotFound desc = could not find container \"d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58\": container with ID starting with d2734321b33876a6ad43bd3ede4966307cac67866be132d267dbdce6950bfb58 not found: ID does not exist" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.483220 4871 scope.go:117] "RemoveContainer" containerID="707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d" Oct 07 23:22:03 crc kubenswrapper[4871]: E1007 23:22:03.483711 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d\": container with ID starting with 707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d not found: ID does not exist" containerID="707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.483758 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d"} err="failed to get container status \"707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d\": rpc error: code = NotFound desc = could not find container \"707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d\": container with ID starting with 707a3fbb285e077041fa3126b3680ca664bcc2afa60470dbe99388648eda294d not found: ID does not exist" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.483810 4871 scope.go:117] "RemoveContainer" containerID="91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5" Oct 07 23:22:03 crc kubenswrapper[4871]: E1007 23:22:03.484196 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5\": container with ID starting with 91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5 not found: ID does not exist" containerID="91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5" Oct 07 23:22:03 crc kubenswrapper[4871]: I1007 23:22:03.484264 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5"} err="failed to get container status \"91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5\": rpc error: code = NotFound desc = could not find container \"91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5\": container with ID starting with 91c2426d849250fbe8afe8171e9f590599f47ed8eef7b04341388e89982bd9d5 not found: ID does not exist" Oct 07 23:22:04 crc kubenswrapper[4871]: I1007 23:22:04.995321 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93a98e09-9cdd-4616-be66-1c362992fb02" path="/var/lib/kubelet/pods/93a98e09-9cdd-4616-be66-1c362992fb02/volumes" Oct 07 23:22:10 crc kubenswrapper[4871]: I1007 23:22:10.983436 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:22:10 crc kubenswrapper[4871]: E1007 23:22:10.984223 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:22:25 crc kubenswrapper[4871]: I1007 23:22:25.982624 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:22:25 crc kubenswrapper[4871]: E1007 23:22:25.983684 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:22:37 crc kubenswrapper[4871]: I1007 23:22:37.983126 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:22:37 crc kubenswrapper[4871]: E1007 23:22:37.983927 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:22:48 crc kubenswrapper[4871]: I1007 23:22:48.982633 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:22:48 crc kubenswrapper[4871]: E1007 23:22:48.983353 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:23:03 crc kubenswrapper[4871]: I1007 23:23:03.982349 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:23:03 crc kubenswrapper[4871]: E1007 23:23:03.984486 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:23:14 crc kubenswrapper[4871]: I1007 23:23:14.982353 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:23:14 crc kubenswrapper[4871]: E1007 23:23:14.983263 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:23:29 crc kubenswrapper[4871]: I1007 23:23:29.982769 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:23:29 crc kubenswrapper[4871]: E1007 23:23:29.983597 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:23:40 crc kubenswrapper[4871]: I1007 23:23:40.983005 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:23:40 crc kubenswrapper[4871]: E1007 23:23:40.984177 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:23:55 crc kubenswrapper[4871]: I1007 23:23:55.588990 4871 scope.go:117] "RemoveContainer" containerID="2e410a288b0644e77d7d50d658c333ceffd37e406e932bcbcbe9a3c3c8c7c744" Oct 07 23:23:55 crc kubenswrapper[4871]: I1007 23:23:55.982120 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:23:55 crc kubenswrapper[4871]: E1007 23:23:55.982480 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:24:10 crc kubenswrapper[4871]: I1007 23:24:10.983093 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:24:10 crc kubenswrapper[4871]: E1007 23:24:10.985672 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:24:23 crc kubenswrapper[4871]: I1007 23:24:23.982565 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:24:23 crc kubenswrapper[4871]: E1007 23:24:23.983680 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:24:32 crc kubenswrapper[4871]: I1007 23:24:32.853721 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-hntcc"] Oct 07 23:24:32 crc kubenswrapper[4871]: I1007 23:24:32.864123 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-hntcc"] Oct 07 23:24:32 crc kubenswrapper[4871]: I1007 23:24:32.996723 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7cd42c4-268b-4588-a427-0a80502fe808" path="/var/lib/kubelet/pods/d7cd42c4-268b-4588-a427-0a80502fe808/volumes" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.005997 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-pzwxv"] Oct 07 23:24:33 crc kubenswrapper[4871]: E1007 23:24:33.006660 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a98e09-9cdd-4616-be66-1c362992fb02" containerName="extract-utilities" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.006710 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a98e09-9cdd-4616-be66-1c362992fb02" containerName="extract-utilities" Oct 07 23:24:33 crc kubenswrapper[4871]: E1007 23:24:33.006752 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a98e09-9cdd-4616-be66-1c362992fb02" containerName="extract-content" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.006771 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a98e09-9cdd-4616-be66-1c362992fb02" containerName="extract-content" Oct 07 23:24:33 crc kubenswrapper[4871]: E1007 23:24:33.006862 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a98e09-9cdd-4616-be66-1c362992fb02" containerName="registry-server" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.006883 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a98e09-9cdd-4616-be66-1c362992fb02" containerName="registry-server" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.007347 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a98e09-9cdd-4616-be66-1c362992fb02" containerName="registry-server" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.010429 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.013745 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.015035 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-pzwxv"] Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.015346 4871 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-595tx" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.015947 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.016358 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.193062 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-crc-storage\") pod \"crc-storage-crc-pzwxv\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.193167 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-node-mnt\") pod \"crc-storage-crc-pzwxv\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.193248 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd8lg\" (UniqueName: \"kubernetes.io/projected/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-kube-api-access-xd8lg\") pod \"crc-storage-crc-pzwxv\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.294553 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-crc-storage\") pod \"crc-storage-crc-pzwxv\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.295088 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-node-mnt\") pod \"crc-storage-crc-pzwxv\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.295129 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd8lg\" (UniqueName: \"kubernetes.io/projected/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-kube-api-access-xd8lg\") pod \"crc-storage-crc-pzwxv\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.295940 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-node-mnt\") pod \"crc-storage-crc-pzwxv\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.296306 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-crc-storage\") pod \"crc-storage-crc-pzwxv\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.329047 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd8lg\" (UniqueName: \"kubernetes.io/projected/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-kube-api-access-xd8lg\") pod \"crc-storage-crc-pzwxv\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.351166 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.865415 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-pzwxv"] Oct 07 23:24:33 crc kubenswrapper[4871]: W1007 23:24:33.870413 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod818fd47a_efab_4a63_bbf1_4c1fa721a2cd.slice/crio-9ea21b0dd3b16774931cc9470b08380c0f9ec99ada3386a28bf97f6e81ffab4d WatchSource:0}: Error finding container 9ea21b0dd3b16774931cc9470b08380c0f9ec99ada3386a28bf97f6e81ffab4d: Status 404 returned error can't find the container with id 9ea21b0dd3b16774931cc9470b08380c0f9ec99ada3386a28bf97f6e81ffab4d Oct 07 23:24:33 crc kubenswrapper[4871]: I1007 23:24:33.876064 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 23:24:34 crc kubenswrapper[4871]: I1007 23:24:34.827950 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pzwxv" event={"ID":"818fd47a-efab-4a63-bbf1-4c1fa721a2cd","Type":"ContainerStarted","Data":"b386ef48a4671a721e35166bc7d37f21fb26ddf6426a8817770002609981cab5"} Oct 07 23:24:34 crc kubenswrapper[4871]: I1007 23:24:34.828031 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pzwxv" event={"ID":"818fd47a-efab-4a63-bbf1-4c1fa721a2cd","Type":"ContainerStarted","Data":"9ea21b0dd3b16774931cc9470b08380c0f9ec99ada3386a28bf97f6e81ffab4d"} Oct 07 23:24:34 crc kubenswrapper[4871]: I1007 23:24:34.861121 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-pzwxv" podStartSLOduration=2.211965369 podStartE2EDuration="2.861097373s" podCreationTimestamp="2025-10-07 23:24:32 +0000 UTC" firstStartedPulling="2025-10-07 23:24:33.875573445 +0000 UTC m=+4547.678271548" lastFinishedPulling="2025-10-07 23:24:34.524705479 +0000 UTC m=+4548.327403552" observedRunningTime="2025-10-07 23:24:34.854277233 +0000 UTC m=+4548.656975376" watchObservedRunningTime="2025-10-07 23:24:34.861097373 +0000 UTC m=+4548.663795486" Oct 07 23:24:35 crc kubenswrapper[4871]: I1007 23:24:35.838658 4871 generic.go:334] "Generic (PLEG): container finished" podID="818fd47a-efab-4a63-bbf1-4c1fa721a2cd" containerID="b386ef48a4671a721e35166bc7d37f21fb26ddf6426a8817770002609981cab5" exitCode=0 Oct 07 23:24:35 crc kubenswrapper[4871]: I1007 23:24:35.839038 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pzwxv" event={"ID":"818fd47a-efab-4a63-bbf1-4c1fa721a2cd","Type":"ContainerDied","Data":"b386ef48a4671a721e35166bc7d37f21fb26ddf6426a8817770002609981cab5"} Oct 07 23:24:35 crc kubenswrapper[4871]: I1007 23:24:35.982852 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:24:35 crc kubenswrapper[4871]: E1007 23:24:35.983848 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.231254 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.363965 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-node-mnt\") pod \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.364046 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-crc-storage\") pod \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.364125 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd8lg\" (UniqueName: \"kubernetes.io/projected/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-kube-api-access-xd8lg\") pod \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\" (UID: \"818fd47a-efab-4a63-bbf1-4c1fa721a2cd\") " Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.364174 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "818fd47a-efab-4a63-bbf1-4c1fa721a2cd" (UID: "818fd47a-efab-4a63-bbf1-4c1fa721a2cd"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.365493 4871 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.374333 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-kube-api-access-xd8lg" (OuterVolumeSpecName: "kube-api-access-xd8lg") pod "818fd47a-efab-4a63-bbf1-4c1fa721a2cd" (UID: "818fd47a-efab-4a63-bbf1-4c1fa721a2cd"). InnerVolumeSpecName "kube-api-access-xd8lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.400387 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "818fd47a-efab-4a63-bbf1-4c1fa721a2cd" (UID: "818fd47a-efab-4a63-bbf1-4c1fa721a2cd"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.467518 4871 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.467576 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd8lg\" (UniqueName: \"kubernetes.io/projected/818fd47a-efab-4a63-bbf1-4c1fa721a2cd-kube-api-access-xd8lg\") on node \"crc\" DevicePath \"\"" Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.862012 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pzwxv" event={"ID":"818fd47a-efab-4a63-bbf1-4c1fa721a2cd","Type":"ContainerDied","Data":"9ea21b0dd3b16774931cc9470b08380c0f9ec99ada3386a28bf97f6e81ffab4d"} Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.862855 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ea21b0dd3b16774931cc9470b08380c0f9ec99ada3386a28bf97f6e81ffab4d" Oct 07 23:24:37 crc kubenswrapper[4871]: I1007 23:24:37.862081 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pzwxv" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.331108 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-pzwxv"] Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.341127 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-pzwxv"] Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.444737 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-fqwdf"] Oct 07 23:24:39 crc kubenswrapper[4871]: E1007 23:24:39.445088 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="818fd47a-efab-4a63-bbf1-4c1fa721a2cd" containerName="storage" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.445104 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="818fd47a-efab-4a63-bbf1-4c1fa721a2cd" containerName="storage" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.445311 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="818fd47a-efab-4a63-bbf1-4c1fa721a2cd" containerName="storage" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.445835 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.449851 4871 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-595tx" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.450574 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.450746 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.451027 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.461477 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-fqwdf"] Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.605029 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4a568e43-3a34-469b-bfde-4833d907ebe8-node-mnt\") pod \"crc-storage-crc-fqwdf\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.605115 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flmwf\" (UniqueName: \"kubernetes.io/projected/4a568e43-3a34-469b-bfde-4833d907ebe8-kube-api-access-flmwf\") pod \"crc-storage-crc-fqwdf\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.605179 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4a568e43-3a34-469b-bfde-4833d907ebe8-crc-storage\") pod \"crc-storage-crc-fqwdf\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.706425 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4a568e43-3a34-469b-bfde-4833d907ebe8-node-mnt\") pod \"crc-storage-crc-fqwdf\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.706518 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flmwf\" (UniqueName: \"kubernetes.io/projected/4a568e43-3a34-469b-bfde-4833d907ebe8-kube-api-access-flmwf\") pod \"crc-storage-crc-fqwdf\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.706578 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4a568e43-3a34-469b-bfde-4833d907ebe8-crc-storage\") pod \"crc-storage-crc-fqwdf\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.707061 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4a568e43-3a34-469b-bfde-4833d907ebe8-node-mnt\") pod \"crc-storage-crc-fqwdf\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.707775 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4a568e43-3a34-469b-bfde-4833d907ebe8-crc-storage\") pod \"crc-storage-crc-fqwdf\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.735918 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flmwf\" (UniqueName: \"kubernetes.io/projected/4a568e43-3a34-469b-bfde-4833d907ebe8-kube-api-access-flmwf\") pod \"crc-storage-crc-fqwdf\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:39 crc kubenswrapper[4871]: I1007 23:24:39.770913 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:40 crc kubenswrapper[4871]: I1007 23:24:40.052503 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-fqwdf"] Oct 07 23:24:40 crc kubenswrapper[4871]: I1007 23:24:40.893970 4871 generic.go:334] "Generic (PLEG): container finished" podID="4a568e43-3a34-469b-bfde-4833d907ebe8" containerID="d34a74a9112997fb3389903d47ea136639dd58b79b8be384c8533b004c5c9c31" exitCode=0 Oct 07 23:24:40 crc kubenswrapper[4871]: I1007 23:24:40.894093 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-fqwdf" event={"ID":"4a568e43-3a34-469b-bfde-4833d907ebe8","Type":"ContainerDied","Data":"d34a74a9112997fb3389903d47ea136639dd58b79b8be384c8533b004c5c9c31"} Oct 07 23:24:40 crc kubenswrapper[4871]: I1007 23:24:40.894546 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-fqwdf" event={"ID":"4a568e43-3a34-469b-bfde-4833d907ebe8","Type":"ContainerStarted","Data":"3e2859f38d8ac5e96519b98dad060bbb75c14cd9b50c3d42dd238ccfe1c0c75a"} Oct 07 23:24:41 crc kubenswrapper[4871]: I1007 23:24:41.006351 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="818fd47a-efab-4a63-bbf1-4c1fa721a2cd" path="/var/lib/kubelet/pods/818fd47a-efab-4a63-bbf1-4c1fa721a2cd/volumes" Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.599914 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.657203 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flmwf\" (UniqueName: \"kubernetes.io/projected/4a568e43-3a34-469b-bfde-4833d907ebe8-kube-api-access-flmwf\") pod \"4a568e43-3a34-469b-bfde-4833d907ebe8\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.657295 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4a568e43-3a34-469b-bfde-4833d907ebe8-node-mnt\") pod \"4a568e43-3a34-469b-bfde-4833d907ebe8\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.657351 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4a568e43-3a34-469b-bfde-4833d907ebe8-crc-storage\") pod \"4a568e43-3a34-469b-bfde-4833d907ebe8\" (UID: \"4a568e43-3a34-469b-bfde-4833d907ebe8\") " Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.657646 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4a568e43-3a34-469b-bfde-4833d907ebe8-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "4a568e43-3a34-469b-bfde-4833d907ebe8" (UID: "4a568e43-3a34-469b-bfde-4833d907ebe8"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.679026 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a568e43-3a34-469b-bfde-4833d907ebe8-kube-api-access-flmwf" (OuterVolumeSpecName: "kube-api-access-flmwf") pod "4a568e43-3a34-469b-bfde-4833d907ebe8" (UID: "4a568e43-3a34-469b-bfde-4833d907ebe8"). InnerVolumeSpecName "kube-api-access-flmwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.692757 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a568e43-3a34-469b-bfde-4833d907ebe8-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "4a568e43-3a34-469b-bfde-4833d907ebe8" (UID: "4a568e43-3a34-469b-bfde-4833d907ebe8"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.758784 4871 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4a568e43-3a34-469b-bfde-4833d907ebe8-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.758852 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flmwf\" (UniqueName: \"kubernetes.io/projected/4a568e43-3a34-469b-bfde-4833d907ebe8-kube-api-access-flmwf\") on node \"crc\" DevicePath \"\"" Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.758868 4871 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4a568e43-3a34-469b-bfde-4833d907ebe8-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.918394 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-fqwdf" event={"ID":"4a568e43-3a34-469b-bfde-4833d907ebe8","Type":"ContainerDied","Data":"3e2859f38d8ac5e96519b98dad060bbb75c14cd9b50c3d42dd238ccfe1c0c75a"} Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.918443 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e2859f38d8ac5e96519b98dad060bbb75c14cd9b50c3d42dd238ccfe1c0c75a" Oct 07 23:24:42 crc kubenswrapper[4871]: I1007 23:24:42.918475 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-fqwdf" Oct 07 23:24:50 crc kubenswrapper[4871]: I1007 23:24:50.983933 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:24:50 crc kubenswrapper[4871]: E1007 23:24:50.985267 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:24:55 crc kubenswrapper[4871]: I1007 23:24:55.638664 4871 scope.go:117] "RemoveContainer" containerID="44944b2b560be5ba8cda32597e753207f059bb65f38c89d93361cb4bc02af202" Oct 07 23:24:55 crc kubenswrapper[4871]: I1007 23:24:55.674564 4871 scope.go:117] "RemoveContainer" containerID="b373140d1a8bd7e0a2a7c1dab5e64d71766cf61ca69697fd4a99ad3ff18304f5" Oct 07 23:24:55 crc kubenswrapper[4871]: I1007 23:24:55.717311 4871 scope.go:117] "RemoveContainer" containerID="23546bfd432fd6f5e99b8618904f405d15eba474ea295730b94ee11ad7d99a04" Oct 07 23:25:01 crc kubenswrapper[4871]: I1007 23:25:01.982453 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:25:01 crc kubenswrapper[4871]: E1007 23:25:01.983721 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:25:15 crc kubenswrapper[4871]: I1007 23:25:15.982311 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:25:15 crc kubenswrapper[4871]: E1007 23:25:15.983475 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:25:27 crc kubenswrapper[4871]: I1007 23:25:27.982752 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:25:27 crc kubenswrapper[4871]: E1007 23:25:27.983660 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:25:40 crc kubenswrapper[4871]: I1007 23:25:40.983127 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:25:41 crc kubenswrapper[4871]: I1007 23:25:41.457427 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"3ab5d945d29b218b1fcfe5ad517fbc83535bc60079d91b4857bc595f89a8605f"} Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.713770 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pw5k8"] Oct 07 23:26:40 crc kubenswrapper[4871]: E1007 23:26:40.714490 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a568e43-3a34-469b-bfde-4833d907ebe8" containerName="storage" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.714501 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a568e43-3a34-469b-bfde-4833d907ebe8" containerName="storage" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.714666 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a568e43-3a34-469b-bfde-4833d907ebe8" containerName="storage" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.715582 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.729085 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pw5k8"] Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.822777 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-catalog-content\") pod \"community-operators-pw5k8\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.822856 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27czg\" (UniqueName: \"kubernetes.io/projected/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-kube-api-access-27czg\") pod \"community-operators-pw5k8\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.822884 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-utilities\") pod \"community-operators-pw5k8\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.924700 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-catalog-content\") pod \"community-operators-pw5k8\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.924762 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27czg\" (UniqueName: \"kubernetes.io/projected/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-kube-api-access-27czg\") pod \"community-operators-pw5k8\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.924808 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-utilities\") pod \"community-operators-pw5k8\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.925441 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-catalog-content\") pod \"community-operators-pw5k8\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.925467 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-utilities\") pod \"community-operators-pw5k8\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:40 crc kubenswrapper[4871]: I1007 23:26:40.948870 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27czg\" (UniqueName: \"kubernetes.io/projected/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-kube-api-access-27czg\") pod \"community-operators-pw5k8\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:41 crc kubenswrapper[4871]: I1007 23:26:41.030713 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:41 crc kubenswrapper[4871]: I1007 23:26:41.524008 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pw5k8"] Oct 07 23:26:42 crc kubenswrapper[4871]: I1007 23:26:42.008422 4871 generic.go:334] "Generic (PLEG): container finished" podID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerID="f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984" exitCode=0 Oct 07 23:26:42 crc kubenswrapper[4871]: I1007 23:26:42.008545 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pw5k8" event={"ID":"2ec84ee8-56b0-4b30-abc3-b713b8b3521e","Type":"ContainerDied","Data":"f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984"} Oct 07 23:26:42 crc kubenswrapper[4871]: I1007 23:26:42.008844 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pw5k8" event={"ID":"2ec84ee8-56b0-4b30-abc3-b713b8b3521e","Type":"ContainerStarted","Data":"a15d05c8ed1a29fc693fa6c4192ba15a16b9e12ac92487a7ec084b339815dbb4"} Oct 07 23:26:44 crc kubenswrapper[4871]: I1007 23:26:44.037916 4871 generic.go:334] "Generic (PLEG): container finished" podID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerID="36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81" exitCode=0 Oct 07 23:26:44 crc kubenswrapper[4871]: I1007 23:26:44.038280 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pw5k8" event={"ID":"2ec84ee8-56b0-4b30-abc3-b713b8b3521e","Type":"ContainerDied","Data":"36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81"} Oct 07 23:26:45 crc kubenswrapper[4871]: I1007 23:26:45.049430 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pw5k8" event={"ID":"2ec84ee8-56b0-4b30-abc3-b713b8b3521e","Type":"ContainerStarted","Data":"ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2"} Oct 07 23:26:45 crc kubenswrapper[4871]: I1007 23:26:45.090176 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pw5k8" podStartSLOduration=2.562839098 podStartE2EDuration="5.090153398s" podCreationTimestamp="2025-10-07 23:26:40 +0000 UTC" firstStartedPulling="2025-10-07 23:26:42.011921936 +0000 UTC m=+4675.814620039" lastFinishedPulling="2025-10-07 23:26:44.539236236 +0000 UTC m=+4678.341934339" observedRunningTime="2025-10-07 23:26:45.084015327 +0000 UTC m=+4678.886713400" watchObservedRunningTime="2025-10-07 23:26:45.090153398 +0000 UTC m=+4678.892851471" Oct 07 23:26:51 crc kubenswrapper[4871]: I1007 23:26:51.031176 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:51 crc kubenswrapper[4871]: I1007 23:26:51.031766 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:51 crc kubenswrapper[4871]: I1007 23:26:51.111336 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:51 crc kubenswrapper[4871]: I1007 23:26:51.179068 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:51 crc kubenswrapper[4871]: I1007 23:26:51.366282 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pw5k8"] Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.122241 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pw5k8" podUID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerName="registry-server" containerID="cri-o://ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2" gracePeriod=2 Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.599042 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.727645 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-utilities\") pod \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.727753 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27czg\" (UniqueName: \"kubernetes.io/projected/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-kube-api-access-27czg\") pod \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.727920 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-catalog-content\") pod \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\" (UID: \"2ec84ee8-56b0-4b30-abc3-b713b8b3521e\") " Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.728870 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-utilities" (OuterVolumeSpecName: "utilities") pod "2ec84ee8-56b0-4b30-abc3-b713b8b3521e" (UID: "2ec84ee8-56b0-4b30-abc3-b713b8b3521e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.737414 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-kube-api-access-27czg" (OuterVolumeSpecName: "kube-api-access-27czg") pod "2ec84ee8-56b0-4b30-abc3-b713b8b3521e" (UID: "2ec84ee8-56b0-4b30-abc3-b713b8b3521e"). InnerVolumeSpecName "kube-api-access-27czg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.787031 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ec84ee8-56b0-4b30-abc3-b713b8b3521e" (UID: "2ec84ee8-56b0-4b30-abc3-b713b8b3521e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.829493 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.829521 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27czg\" (UniqueName: \"kubernetes.io/projected/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-kube-api-access-27czg\") on node \"crc\" DevicePath \"\"" Oct 07 23:26:53 crc kubenswrapper[4871]: I1007 23:26:53.829531 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec84ee8-56b0-4b30-abc3-b713b8b3521e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.130682 4871 generic.go:334] "Generic (PLEG): container finished" podID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerID="ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2" exitCode=0 Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.130738 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pw5k8" event={"ID":"2ec84ee8-56b0-4b30-abc3-b713b8b3521e","Type":"ContainerDied","Data":"ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2"} Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.130776 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pw5k8" event={"ID":"2ec84ee8-56b0-4b30-abc3-b713b8b3521e","Type":"ContainerDied","Data":"a15d05c8ed1a29fc693fa6c4192ba15a16b9e12ac92487a7ec084b339815dbb4"} Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.130833 4871 scope.go:117] "RemoveContainer" containerID="ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.130998 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pw5k8" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.157783 4871 scope.go:117] "RemoveContainer" containerID="36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.170514 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pw5k8"] Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.184950 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pw5k8"] Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.201668 4871 scope.go:117] "RemoveContainer" containerID="f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.222994 4871 scope.go:117] "RemoveContainer" containerID="ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2" Oct 07 23:26:54 crc kubenswrapper[4871]: E1007 23:26:54.223399 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2\": container with ID starting with ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2 not found: ID does not exist" containerID="ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.223442 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2"} err="failed to get container status \"ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2\": rpc error: code = NotFound desc = could not find container \"ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2\": container with ID starting with ec6452c17dc72096e386bb66586ffcb874b066f68681302063f4b60a20a7d8d2 not found: ID does not exist" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.223469 4871 scope.go:117] "RemoveContainer" containerID="36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81" Oct 07 23:26:54 crc kubenswrapper[4871]: E1007 23:26:54.223958 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81\": container with ID starting with 36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81 not found: ID does not exist" containerID="36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.224009 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81"} err="failed to get container status \"36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81\": rpc error: code = NotFound desc = could not find container \"36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81\": container with ID starting with 36fd74459aaf5824289bfdc1fbe92bad8b8e661007a98b89c5e8f99a582c2c81 not found: ID does not exist" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.224032 4871 scope.go:117] "RemoveContainer" containerID="f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984" Oct 07 23:26:54 crc kubenswrapper[4871]: E1007 23:26:54.224361 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984\": container with ID starting with f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984 not found: ID does not exist" containerID="f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.224387 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984"} err="failed to get container status \"f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984\": rpc error: code = NotFound desc = could not find container \"f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984\": container with ID starting with f8ced7943d8abd15e6d78f9b997e84105b9ed6ef7f91d058f37f1ba04e173984 not found: ID does not exist" Oct 07 23:26:54 crc kubenswrapper[4871]: I1007 23:26:54.996233 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" path="/var/lib/kubelet/pods/2ec84ee8-56b0-4b30-abc3-b713b8b3521e/volumes" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.402597 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-6chmr"] Oct 07 23:27:48 crc kubenswrapper[4871]: E1007 23:27:48.403469 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerName="extract-content" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.403483 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerName="extract-content" Oct 07 23:27:48 crc kubenswrapper[4871]: E1007 23:27:48.403504 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerName="registry-server" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.403510 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerName="registry-server" Oct 07 23:27:48 crc kubenswrapper[4871]: E1007 23:27:48.403540 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerName="extract-utilities" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.403547 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerName="extract-utilities" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.403678 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ec84ee8-56b0-4b30-abc3-b713b8b3521e" containerName="registry-server" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.404524 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.407605 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.407764 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.408027 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.408228 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.409256 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-6chmr"] Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.409632 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-jlpg6" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.477173 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92wnx\" (UniqueName: \"kubernetes.io/projected/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-kube-api-access-92wnx\") pod \"dnsmasq-dns-85f98b87f9-6chmr\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.477243 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-config\") pod \"dnsmasq-dns-85f98b87f9-6chmr\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.477300 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-dns-svc\") pod \"dnsmasq-dns-85f98b87f9-6chmr\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.578457 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-config\") pod \"dnsmasq-dns-85f98b87f9-6chmr\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.578545 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-dns-svc\") pod \"dnsmasq-dns-85f98b87f9-6chmr\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.578660 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92wnx\" (UniqueName: \"kubernetes.io/projected/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-kube-api-access-92wnx\") pod \"dnsmasq-dns-85f98b87f9-6chmr\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.579454 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-config\") pod \"dnsmasq-dns-85f98b87f9-6chmr\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.579476 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-dns-svc\") pod \"dnsmasq-dns-85f98b87f9-6chmr\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.607632 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92wnx\" (UniqueName: \"kubernetes.io/projected/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-kube-api-access-92wnx\") pod \"dnsmasq-dns-85f98b87f9-6chmr\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.681443 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-j7kwj"] Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.685098 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.690658 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-j7kwj"] Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.723139 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.781370 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-config\") pod \"dnsmasq-dns-67d9f7fb89-j7kwj\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.781423 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-dns-svc\") pod \"dnsmasq-dns-67d9f7fb89-j7kwj\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.781446 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbh9m\" (UniqueName: \"kubernetes.io/projected/e19bcc51-e926-424e-aecf-b6985e55462e-kube-api-access-sbh9m\") pod \"dnsmasq-dns-67d9f7fb89-j7kwj\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.885561 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-config\") pod \"dnsmasq-dns-67d9f7fb89-j7kwj\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.885648 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-dns-svc\") pod \"dnsmasq-dns-67d9f7fb89-j7kwj\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.885830 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbh9m\" (UniqueName: \"kubernetes.io/projected/e19bcc51-e926-424e-aecf-b6985e55462e-kube-api-access-sbh9m\") pod \"dnsmasq-dns-67d9f7fb89-j7kwj\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.889094 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-config\") pod \"dnsmasq-dns-67d9f7fb89-j7kwj\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.889289 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-dns-svc\") pod \"dnsmasq-dns-67d9f7fb89-j7kwj\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:48 crc kubenswrapper[4871]: I1007 23:27:48.907962 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbh9m\" (UniqueName: \"kubernetes.io/projected/e19bcc51-e926-424e-aecf-b6985e55462e-kube-api-access-sbh9m\") pod \"dnsmasq-dns-67d9f7fb89-j7kwj\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.003018 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.160105 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-6chmr"] Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.240750 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-j7kwj"] Oct 07 23:27:49 crc kubenswrapper[4871]: W1007 23:27:49.246374 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode19bcc51_e926_424e_aecf_b6985e55462e.slice/crio-77e49fa68190e6b5ef5ff4b46270deaa9c939489989f7b2a3bee07ab6affabf7 WatchSource:0}: Error finding container 77e49fa68190e6b5ef5ff4b46270deaa9c939489989f7b2a3bee07ab6affabf7: Status 404 returned error can't find the container with id 77e49fa68190e6b5ef5ff4b46270deaa9c939489989f7b2a3bee07ab6affabf7 Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.569488 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.571629 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.583442 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.584846 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.585045 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.585087 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.585295 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.585745 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-qk4qd" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.659209 4871 generic.go:334] "Generic (PLEG): container finished" podID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" containerID="63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4" exitCode=0 Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.659303 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" event={"ID":"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4","Type":"ContainerDied","Data":"63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4"} Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.659639 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" event={"ID":"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4","Type":"ContainerStarted","Data":"39e65e127ff18a7d293ed3217f54c24c5e9bd56fce23e7f89653c0aa27dbdeeb"} Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.661549 4871 generic.go:334] "Generic (PLEG): container finished" podID="e19bcc51-e926-424e-aecf-b6985e55462e" containerID="37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b" exitCode=0 Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.661587 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" event={"ID":"e19bcc51-e926-424e-aecf-b6985e55462e","Type":"ContainerDied","Data":"37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b"} Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.661610 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" event={"ID":"e19bcc51-e926-424e-aecf-b6985e55462e","Type":"ContainerStarted","Data":"77e49fa68190e6b5ef5ff4b46270deaa9c939489989f7b2a3bee07ab6affabf7"} Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.701087 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.701156 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-server-conf\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.701179 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9ft8\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-kube-api-access-t9ft8\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.701216 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.701289 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.701315 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-pod-info\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.701340 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.701367 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.701413 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.803025 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9ft8\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-kube-api-access-t9ft8\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.803159 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.803328 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.803381 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-pod-info\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.803428 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.803479 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.803575 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.803681 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.803727 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-server-conf\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.804108 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.805017 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.805169 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-server-conf\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.806412 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.809893 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.810135 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.810232 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.810335 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2c8faa66fce466b48b425a0e4cbd921197e1b803b01adab0d2859d6a441f6bef/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.812168 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-pod-info\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.827683 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.828804 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.828935 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9ft8\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-kube-api-access-t9ft8\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.830729 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-w52c2" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.833393 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.833409 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.833632 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.834112 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.846578 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 23:27:49 crc kubenswrapper[4871]: E1007 23:27:49.865377 4871 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 07 23:27:49 crc kubenswrapper[4871]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 07 23:27:49 crc kubenswrapper[4871]: > podSandboxID="39e65e127ff18a7d293ed3217f54c24c5e9bd56fce23e7f89653c0aa27dbdeeb" Oct 07 23:27:49 crc kubenswrapper[4871]: E1007 23:27:49.865529 4871 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 07 23:27:49 crc kubenswrapper[4871]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-92wnx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-85f98b87f9-6chmr_openstack(52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 07 23:27:49 crc kubenswrapper[4871]: > logger="UnhandledError" Oct 07 23:27:49 crc kubenswrapper[4871]: E1007 23:27:49.870097 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" podUID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.883728 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") pod \"rabbitmq-server-0\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " pod="openstack/rabbitmq-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.905389 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.905446 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7294a0d0-936c-4501-9afe-9ab306906ad2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.905480 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.905511 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.905548 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.905571 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.905658 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52nk6\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-kube-api-access-52nk6\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.905690 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7294a0d0-936c-4501-9afe-9ab306906ad2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.905760 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:49 crc kubenswrapper[4871]: I1007 23:27:49.964292 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.007329 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.007398 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52nk6\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-kube-api-access-52nk6\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.007443 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7294a0d0-936c-4501-9afe-9ab306906ad2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.007478 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.007505 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.007550 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7294a0d0-936c-4501-9afe-9ab306906ad2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.007580 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.007614 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.007665 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.008005 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.008674 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.009939 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.010697 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.011408 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.011438 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a649604412d27923c0fe096e2ba7c6ce60c40b8e70d73cbc9cac8dd896ba0a3c/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.013696 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.014034 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7294a0d0-936c-4501-9afe-9ab306906ad2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.015343 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7294a0d0-936c-4501-9afe-9ab306906ad2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.032295 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52nk6\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-kube-api-access-52nk6\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.056851 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") pod \"rabbitmq-cell1-server-0\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.148591 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.213507 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.606975 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 23:27:50 crc kubenswrapper[4871]: W1007 23:27:50.612138 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7294a0d0_936c_4501_9afe_9ab306906ad2.slice/crio-22e13d38d8572dfbb80aa459e04eeea4b464ef8bc97202891e45f0382459a3b7 WatchSource:0}: Error finding container 22e13d38d8572dfbb80aa459e04eeea4b464ef8bc97202891e45f0382459a3b7: Status 404 returned error can't find the container with id 22e13d38d8572dfbb80aa459e04eeea4b464ef8bc97202891e45f0382459a3b7 Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.693005 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" event={"ID":"e19bcc51-e926-424e-aecf-b6985e55462e","Type":"ContainerStarted","Data":"d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f"} Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.701019 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.714114 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"116a6f25-47bf-4d65-aaf8-d3efe5bc8162","Type":"ContainerStarted","Data":"e7baa8626363e1db9c16fee18e8b24524795e24fc21eeed12a71e26782d2ea5a"} Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.719234 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7294a0d0-936c-4501-9afe-9ab306906ad2","Type":"ContainerStarted","Data":"22e13d38d8572dfbb80aa459e04eeea4b464ef8bc97202891e45f0382459a3b7"} Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.731840 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" podStartSLOduration=2.73181982 podStartE2EDuration="2.73181982s" podCreationTimestamp="2025-10-07 23:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:27:50.725476923 +0000 UTC m=+4744.528175006" watchObservedRunningTime="2025-10-07 23:27:50.73181982 +0000 UTC m=+4744.534517913" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.934605 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.935457 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.937318 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.938785 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-24kqg" Oct 07 23:27:50 crc kubenswrapper[4871]: I1007 23:27:50.948885 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.023865 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cd8j\" (UniqueName: \"kubernetes.io/projected/2985dd37-b0b5-4d84-9a26-cddfaf27be64-kube-api-access-8cd8j\") pod \"memcached-0\" (UID: \"2985dd37-b0b5-4d84-9a26-cddfaf27be64\") " pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.025078 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2985dd37-b0b5-4d84-9a26-cddfaf27be64-config-data\") pod \"memcached-0\" (UID: \"2985dd37-b0b5-4d84-9a26-cddfaf27be64\") " pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.025229 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2985dd37-b0b5-4d84-9a26-cddfaf27be64-kolla-config\") pod \"memcached-0\" (UID: \"2985dd37-b0b5-4d84-9a26-cddfaf27be64\") " pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.127066 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2985dd37-b0b5-4d84-9a26-cddfaf27be64-kolla-config\") pod \"memcached-0\" (UID: \"2985dd37-b0b5-4d84-9a26-cddfaf27be64\") " pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.127137 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cd8j\" (UniqueName: \"kubernetes.io/projected/2985dd37-b0b5-4d84-9a26-cddfaf27be64-kube-api-access-8cd8j\") pod \"memcached-0\" (UID: \"2985dd37-b0b5-4d84-9a26-cddfaf27be64\") " pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.127193 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2985dd37-b0b5-4d84-9a26-cddfaf27be64-config-data\") pod \"memcached-0\" (UID: \"2985dd37-b0b5-4d84-9a26-cddfaf27be64\") " pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.127972 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2985dd37-b0b5-4d84-9a26-cddfaf27be64-config-data\") pod \"memcached-0\" (UID: \"2985dd37-b0b5-4d84-9a26-cddfaf27be64\") " pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.128352 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2985dd37-b0b5-4d84-9a26-cddfaf27be64-kolla-config\") pod \"memcached-0\" (UID: \"2985dd37-b0b5-4d84-9a26-cddfaf27be64\") " pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.277383 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cd8j\" (UniqueName: \"kubernetes.io/projected/2985dd37-b0b5-4d84-9a26-cddfaf27be64-kube-api-access-8cd8j\") pod \"memcached-0\" (UID: \"2985dd37-b0b5-4d84-9a26-cddfaf27be64\") " pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.376970 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.730666 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" event={"ID":"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4","Type":"ContainerStarted","Data":"0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be"} Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.731240 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.732678 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"116a6f25-47bf-4d65-aaf8-d3efe5bc8162","Type":"ContainerStarted","Data":"648fc38c263a9fdb4358e253bd667edffdaa9b9f99c5658764ca1f46e6ea0666"} Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.759883 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" podStartSLOduration=3.75986051 podStartE2EDuration="3.75986051s" podCreationTimestamp="2025-10-07 23:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:27:51.753964344 +0000 UTC m=+4745.556662457" watchObservedRunningTime="2025-10-07 23:27:51.75986051 +0000 UTC m=+4745.562558623" Oct 07 23:27:51 crc kubenswrapper[4871]: I1007 23:27:51.813171 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.269362 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.270578 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.274766 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-c2hw6" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.274883 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.275691 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.275692 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.276216 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.293261 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.300983 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.317880 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.329257 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.332471 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.332770 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.333425 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-wjg5f" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.334081 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.343627 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346141 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f780ea63-c7a1-4fdb-828d-bf2c1492e300-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346197 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drnwc\" (UniqueName: \"kubernetes.io/projected/f780ea63-c7a1-4fdb-828d-bf2c1492e300-kube-api-access-drnwc\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346250 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b15d0939-9849-40a0-a164-815f7c0baf07\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b15d0939-9849-40a0-a164-815f7c0baf07\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346296 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f780ea63-c7a1-4fdb-828d-bf2c1492e300-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346335 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4b8bd380-bfd3-4607-8a21-3b36dae4a056\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b8bd380-bfd3-4607-8a21-3b36dae4a056\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346369 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/08850732-6460-4677-a906-b8d932f2f1ec-secrets\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346402 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f780ea63-c7a1-4fdb-828d-bf2c1492e300-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346442 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f780ea63-c7a1-4fdb-828d-bf2c1492e300-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346491 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/08850732-6460-4677-a906-b8d932f2f1ec-config-data-generated\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346534 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/08850732-6460-4677-a906-b8d932f2f1ec-config-data-default\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346568 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nlv7\" (UniqueName: \"kubernetes.io/projected/08850732-6460-4677-a906-b8d932f2f1ec-kube-api-access-4nlv7\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346636 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08850732-6460-4677-a906-b8d932f2f1ec-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346671 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f780ea63-c7a1-4fdb-828d-bf2c1492e300-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346736 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f780ea63-c7a1-4fdb-828d-bf2c1492e300-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346781 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f780ea63-c7a1-4fdb-828d-bf2c1492e300-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346847 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08850732-6460-4677-a906-b8d932f2f1ec-operator-scripts\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346923 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/08850732-6460-4677-a906-b8d932f2f1ec-kolla-config\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.346959 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/08850732-6460-4677-a906-b8d932f2f1ec-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448591 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/08850732-6460-4677-a906-b8d932f2f1ec-kolla-config\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448634 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/08850732-6460-4677-a906-b8d932f2f1ec-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448682 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f780ea63-c7a1-4fdb-828d-bf2c1492e300-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448700 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drnwc\" (UniqueName: \"kubernetes.io/projected/f780ea63-c7a1-4fdb-828d-bf2c1492e300-kube-api-access-drnwc\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448727 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b15d0939-9849-40a0-a164-815f7c0baf07\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b15d0939-9849-40a0-a164-815f7c0baf07\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448755 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f780ea63-c7a1-4fdb-828d-bf2c1492e300-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448778 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4b8bd380-bfd3-4607-8a21-3b36dae4a056\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b8bd380-bfd3-4607-8a21-3b36dae4a056\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448811 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/08850732-6460-4677-a906-b8d932f2f1ec-secrets\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448831 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f780ea63-c7a1-4fdb-828d-bf2c1492e300-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448857 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f780ea63-c7a1-4fdb-828d-bf2c1492e300-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448893 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/08850732-6460-4677-a906-b8d932f2f1ec-config-data-generated\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448924 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/08850732-6460-4677-a906-b8d932f2f1ec-config-data-default\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448940 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nlv7\" (UniqueName: \"kubernetes.io/projected/08850732-6460-4677-a906-b8d932f2f1ec-kube-api-access-4nlv7\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.448990 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08850732-6460-4677-a906-b8d932f2f1ec-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.449012 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f780ea63-c7a1-4fdb-828d-bf2c1492e300-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.449061 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f780ea63-c7a1-4fdb-828d-bf2c1492e300-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.449079 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f780ea63-c7a1-4fdb-828d-bf2c1492e300-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.449097 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08850732-6460-4677-a906-b8d932f2f1ec-operator-scripts\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.449405 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f780ea63-c7a1-4fdb-828d-bf2c1492e300-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.449969 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/08850732-6460-4677-a906-b8d932f2f1ec-kolla-config\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.451008 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f780ea63-c7a1-4fdb-828d-bf2c1492e300-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.451009 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f780ea63-c7a1-4fdb-828d-bf2c1492e300-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.452255 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/08850732-6460-4677-a906-b8d932f2f1ec-config-data-generated\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.453440 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/08850732-6460-4677-a906-b8d932f2f1ec-config-data-default\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.453602 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08850732-6460-4677-a906-b8d932f2f1ec-operator-scripts\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.455351 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f780ea63-c7a1-4fdb-828d-bf2c1492e300-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.456201 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.456257 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4b8bd380-bfd3-4607-8a21-3b36dae4a056\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b8bd380-bfd3-4607-8a21-3b36dae4a056\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/31192d1105cf0a3fd44e8bb28fa2b6e720255a31af920745c3452dbe07c483fe/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.457506 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/08850732-6460-4677-a906-b8d932f2f1ec-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.458224 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08850732-6460-4677-a906-b8d932f2f1ec-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.458627 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f780ea63-c7a1-4fdb-828d-bf2c1492e300-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.461420 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/08850732-6460-4677-a906-b8d932f2f1ec-secrets\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.462766 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.462868 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b15d0939-9849-40a0-a164-815f7c0baf07\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b15d0939-9849-40a0-a164-815f7c0baf07\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/77bfe8a64d4f5de9400736ca761cedc18c7a749cad89776d056c2db8eede44c6/globalmount\"" pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.467897 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f780ea63-c7a1-4fdb-828d-bf2c1492e300-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.472413 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drnwc\" (UniqueName: \"kubernetes.io/projected/f780ea63-c7a1-4fdb-828d-bf2c1492e300-kube-api-access-drnwc\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.473241 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f780ea63-c7a1-4fdb-828d-bf2c1492e300-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.480776 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nlv7\" (UniqueName: \"kubernetes.io/projected/08850732-6460-4677-a906-b8d932f2f1ec-kube-api-access-4nlv7\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.504454 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4b8bd380-bfd3-4607-8a21-3b36dae4a056\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b8bd380-bfd3-4607-8a21-3b36dae4a056\") pod \"openstack-cell1-galera-0\" (UID: \"f780ea63-c7a1-4fdb-828d-bf2c1492e300\") " pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.505806 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b15d0939-9849-40a0-a164-815f7c0baf07\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b15d0939-9849-40a0-a164-815f7c0baf07\") pod \"openstack-galera-0\" (UID: \"08850732-6460-4677-a906-b8d932f2f1ec\") " pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.610299 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.649213 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.745064 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2985dd37-b0b5-4d84-9a26-cddfaf27be64","Type":"ContainerStarted","Data":"ff85ba33eb16e1c8e86dd2f6b3f4ffb6af3846e4f85f917a33af8d7f13d83fdf"} Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.745129 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2985dd37-b0b5-4d84-9a26-cddfaf27be64","Type":"ContainerStarted","Data":"9e82d1b50cf06b8a33d58832d21eef223dd5f6254941d8d420a03bb38778d4d3"} Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.745168 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.749348 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7294a0d0-936c-4501-9afe-9ab306906ad2","Type":"ContainerStarted","Data":"a5288a75ee8a01032fe8be52619292a8b1261515d2c65c655e9b313a4c3e34c3"} Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.799184 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.799155856 podStartE2EDuration="2.799155856s" podCreationTimestamp="2025-10-07 23:27:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:27:52.770401317 +0000 UTC m=+4746.573099410" watchObservedRunningTime="2025-10-07 23:27:52.799155856 +0000 UTC m=+4746.601853959" Oct 07 23:27:52 crc kubenswrapper[4871]: I1007 23:27:52.980694 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 23:27:52 crc kubenswrapper[4871]: W1007 23:27:52.988640 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08850732_6460_4677_a906_b8d932f2f1ec.slice/crio-077bab876073a38421d7906d974a304f93a9bb106f0056f1fa594c815b438abe WatchSource:0}: Error finding container 077bab876073a38421d7906d974a304f93a9bb106f0056f1fa594c815b438abe: Status 404 returned error can't find the container with id 077bab876073a38421d7906d974a304f93a9bb106f0056f1fa594c815b438abe Oct 07 23:27:53 crc kubenswrapper[4871]: I1007 23:27:53.105745 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 23:27:53 crc kubenswrapper[4871]: W1007 23:27:53.111324 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf780ea63_c7a1_4fdb_828d_bf2c1492e300.slice/crio-878faac1046150bb1ececeddfcf89a0691cfcdb4cfd2daa10117d3a473151c05 WatchSource:0}: Error finding container 878faac1046150bb1ececeddfcf89a0691cfcdb4cfd2daa10117d3a473151c05: Status 404 returned error can't find the container with id 878faac1046150bb1ececeddfcf89a0691cfcdb4cfd2daa10117d3a473151c05 Oct 07 23:27:53 crc kubenswrapper[4871]: I1007 23:27:53.761984 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f780ea63-c7a1-4fdb-828d-bf2c1492e300","Type":"ContainerStarted","Data":"89cb9ef680e7dc8337304968f245bbe4369f68c5aef3fe52cedab06913057af7"} Oct 07 23:27:53 crc kubenswrapper[4871]: I1007 23:27:53.762597 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f780ea63-c7a1-4fdb-828d-bf2c1492e300","Type":"ContainerStarted","Data":"878faac1046150bb1ececeddfcf89a0691cfcdb4cfd2daa10117d3a473151c05"} Oct 07 23:27:53 crc kubenswrapper[4871]: I1007 23:27:53.763990 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"08850732-6460-4677-a906-b8d932f2f1ec","Type":"ContainerStarted","Data":"946a314312094bbc10188886619d07e83b3da4f8595d798586432216e373b3ff"} Oct 07 23:27:53 crc kubenswrapper[4871]: I1007 23:27:53.764051 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"08850732-6460-4677-a906-b8d932f2f1ec","Type":"ContainerStarted","Data":"077bab876073a38421d7906d974a304f93a9bb106f0056f1fa594c815b438abe"} Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.003961 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.052255 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-6chmr"] Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.052465 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" podUID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" containerName="dnsmasq-dns" containerID="cri-o://0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be" gracePeriod=10 Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.536172 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.694604 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-config\") pod \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.694648 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92wnx\" (UniqueName: \"kubernetes.io/projected/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-kube-api-access-92wnx\") pod \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.694780 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-dns-svc\") pod \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\" (UID: \"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4\") " Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.707362 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-kube-api-access-92wnx" (OuterVolumeSpecName: "kube-api-access-92wnx") pod "52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" (UID: "52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4"). InnerVolumeSpecName "kube-api-access-92wnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.739383 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" (UID: "52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.740056 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-config" (OuterVolumeSpecName: "config") pod "52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" (UID: "52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.780109 4871 generic.go:334] "Generic (PLEG): container finished" podID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" containerID="0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be" exitCode=0 Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.780171 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.780188 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" event={"ID":"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4","Type":"ContainerDied","Data":"0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be"} Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.781072 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-6chmr" event={"ID":"52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4","Type":"ContainerDied","Data":"39e65e127ff18a7d293ed3217f54c24c5e9bd56fce23e7f89653c0aa27dbdeeb"} Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.781092 4871 scope.go:117] "RemoveContainer" containerID="0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.796625 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.796677 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.796695 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92wnx\" (UniqueName: \"kubernetes.io/projected/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4-kube-api-access-92wnx\") on node \"crc\" DevicePath \"\"" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.801295 4871 scope.go:117] "RemoveContainer" containerID="63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.819400 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-6chmr"] Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.825290 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-6chmr"] Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.847146 4871 scope.go:117] "RemoveContainer" containerID="0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be" Oct 07 23:27:54 crc kubenswrapper[4871]: E1007 23:27:54.847449 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be\": container with ID starting with 0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be not found: ID does not exist" containerID="0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.847477 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be"} err="failed to get container status \"0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be\": rpc error: code = NotFound desc = could not find container \"0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be\": container with ID starting with 0968f57bddd93aa2af31df7ca75674aa970be382f98e5dfe0257b25f5948a7be not found: ID does not exist" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.847495 4871 scope.go:117] "RemoveContainer" containerID="63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4" Oct 07 23:27:54 crc kubenswrapper[4871]: E1007 23:27:54.847826 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4\": container with ID starting with 63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4 not found: ID does not exist" containerID="63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.847845 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4"} err="failed to get container status \"63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4\": rpc error: code = NotFound desc = could not find container \"63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4\": container with ID starting with 63155c43d747fa4292875331aa9b697792d387f6b856800cc7884f9733b1a6e4 not found: ID does not exist" Oct 07 23:27:54 crc kubenswrapper[4871]: I1007 23:27:54.993011 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" path="/var/lib/kubelet/pods/52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4/volumes" Oct 07 23:27:57 crc kubenswrapper[4871]: I1007 23:27:57.816966 4871 generic.go:334] "Generic (PLEG): container finished" podID="f780ea63-c7a1-4fdb-828d-bf2c1492e300" containerID="89cb9ef680e7dc8337304968f245bbe4369f68c5aef3fe52cedab06913057af7" exitCode=0 Oct 07 23:27:57 crc kubenswrapper[4871]: I1007 23:27:57.817092 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f780ea63-c7a1-4fdb-828d-bf2c1492e300","Type":"ContainerDied","Data":"89cb9ef680e7dc8337304968f245bbe4369f68c5aef3fe52cedab06913057af7"} Oct 07 23:27:57 crc kubenswrapper[4871]: I1007 23:27:57.820783 4871 generic.go:334] "Generic (PLEG): container finished" podID="08850732-6460-4677-a906-b8d932f2f1ec" containerID="946a314312094bbc10188886619d07e83b3da4f8595d798586432216e373b3ff" exitCode=0 Oct 07 23:27:57 crc kubenswrapper[4871]: I1007 23:27:57.820892 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"08850732-6460-4677-a906-b8d932f2f1ec","Type":"ContainerDied","Data":"946a314312094bbc10188886619d07e83b3da4f8595d798586432216e373b3ff"} Oct 07 23:27:58 crc kubenswrapper[4871]: I1007 23:27:58.835644 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f780ea63-c7a1-4fdb-828d-bf2c1492e300","Type":"ContainerStarted","Data":"f788b5cc8e516879db43db8546dc7906026b03343b30b0578e7165e1cdf7d283"} Oct 07 23:27:58 crc kubenswrapper[4871]: I1007 23:27:58.843978 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"08850732-6460-4677-a906-b8d932f2f1ec","Type":"ContainerStarted","Data":"2720d0aabff6df574209f2fcaa335b508d0f088078d135a825597a5e92bcc234"} Oct 07 23:27:58 crc kubenswrapper[4871]: I1007 23:27:58.877849 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.87777968 podStartE2EDuration="7.87777968s" podCreationTimestamp="2025-10-07 23:27:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:27:58.868678339 +0000 UTC m=+4752.671376452" watchObservedRunningTime="2025-10-07 23:27:58.87777968 +0000 UTC m=+4752.680477793" Oct 07 23:27:58 crc kubenswrapper[4871]: I1007 23:27:58.928497 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.928480377 podStartE2EDuration="7.928480377s" podCreationTimestamp="2025-10-07 23:27:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:27:58.926743541 +0000 UTC m=+4752.729441614" watchObservedRunningTime="2025-10-07 23:27:58.928480377 +0000 UTC m=+4752.731178450" Oct 07 23:28:01 crc kubenswrapper[4871]: I1007 23:28:01.379136 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 07 23:28:02 crc kubenswrapper[4871]: I1007 23:28:02.610899 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 07 23:28:02 crc kubenswrapper[4871]: I1007 23:28:02.611354 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 07 23:28:02 crc kubenswrapper[4871]: I1007 23:28:02.650914 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 07 23:28:02 crc kubenswrapper[4871]: I1007 23:28:02.651406 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 07 23:28:02 crc kubenswrapper[4871]: I1007 23:28:02.672009 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 07 23:28:02 crc kubenswrapper[4871]: I1007 23:28:02.957074 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 07 23:28:04 crc kubenswrapper[4871]: I1007 23:28:04.731753 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 07 23:28:04 crc kubenswrapper[4871]: I1007 23:28:04.811873 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 07 23:28:05 crc kubenswrapper[4871]: I1007 23:28:05.512578 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:28:05 crc kubenswrapper[4871]: I1007 23:28:05.512675 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.426372 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2ds8g"] Oct 07 23:28:08 crc kubenswrapper[4871]: E1007 23:28:08.426957 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" containerName="dnsmasq-dns" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.426969 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" containerName="dnsmasq-dns" Oct 07 23:28:08 crc kubenswrapper[4871]: E1007 23:28:08.426986 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" containerName="init" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.426992 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" containerName="init" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.427130 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="52d4a81e-6c4b-4f64-a5ae-0e51c84e71e4" containerName="dnsmasq-dns" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.428058 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.443199 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2ds8g"] Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.538180 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-catalog-content\") pod \"redhat-operators-2ds8g\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.538558 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-utilities\") pod \"redhat-operators-2ds8g\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.538724 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjg7k\" (UniqueName: \"kubernetes.io/projected/130495b5-b37a-4e34-ad8d-b867ee3838f9-kube-api-access-qjg7k\") pod \"redhat-operators-2ds8g\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.640249 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-utilities\") pod \"redhat-operators-2ds8g\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.640716 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjg7k\" (UniqueName: \"kubernetes.io/projected/130495b5-b37a-4e34-ad8d-b867ee3838f9-kube-api-access-qjg7k\") pod \"redhat-operators-2ds8g\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.641009 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-catalog-content\") pod \"redhat-operators-2ds8g\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.641138 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-utilities\") pod \"redhat-operators-2ds8g\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.641664 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-catalog-content\") pod \"redhat-operators-2ds8g\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.667607 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjg7k\" (UniqueName: \"kubernetes.io/projected/130495b5-b37a-4e34-ad8d-b867ee3838f9-kube-api-access-qjg7k\") pod \"redhat-operators-2ds8g\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:08 crc kubenswrapper[4871]: I1007 23:28:08.752187 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:09 crc kubenswrapper[4871]: I1007 23:28:09.086028 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2ds8g"] Oct 07 23:28:09 crc kubenswrapper[4871]: I1007 23:28:09.951335 4871 generic.go:334] "Generic (PLEG): container finished" podID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerID="787ce125fd21c4f80f1cec52d7862b9d28e91647a285000e0cb5a7aa35118a12" exitCode=0 Oct 07 23:28:09 crc kubenswrapper[4871]: I1007 23:28:09.951633 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ds8g" event={"ID":"130495b5-b37a-4e34-ad8d-b867ee3838f9","Type":"ContainerDied","Data":"787ce125fd21c4f80f1cec52d7862b9d28e91647a285000e0cb5a7aa35118a12"} Oct 07 23:28:09 crc kubenswrapper[4871]: I1007 23:28:09.951689 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ds8g" event={"ID":"130495b5-b37a-4e34-ad8d-b867ee3838f9","Type":"ContainerStarted","Data":"b6822a9481b6a423dd6665c0a5e0f367f0932819bce04a1379ea62361161236e"} Oct 07 23:28:10 crc kubenswrapper[4871]: I1007 23:28:10.968158 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ds8g" event={"ID":"130495b5-b37a-4e34-ad8d-b867ee3838f9","Type":"ContainerStarted","Data":"9405df51dcb2c3792b3e3ab7a0987a7a9c5f649ce8734b90f2274fb13a390cd3"} Oct 07 23:28:11 crc kubenswrapper[4871]: I1007 23:28:11.981737 4871 generic.go:334] "Generic (PLEG): container finished" podID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerID="9405df51dcb2c3792b3e3ab7a0987a7a9c5f649ce8734b90f2274fb13a390cd3" exitCode=0 Oct 07 23:28:11 crc kubenswrapper[4871]: I1007 23:28:11.981869 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ds8g" event={"ID":"130495b5-b37a-4e34-ad8d-b867ee3838f9","Type":"ContainerDied","Data":"9405df51dcb2c3792b3e3ab7a0987a7a9c5f649ce8734b90f2274fb13a390cd3"} Oct 07 23:28:12 crc kubenswrapper[4871]: I1007 23:28:12.996784 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ds8g" event={"ID":"130495b5-b37a-4e34-ad8d-b867ee3838f9","Type":"ContainerStarted","Data":"3d3dad8bd943cffdec74358f92205f5ef6e3354d6f169696b22c4f86162de8bb"} Oct 07 23:28:13 crc kubenswrapper[4871]: I1007 23:28:13.015946 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2ds8g" podStartSLOduration=2.589209625 podStartE2EDuration="5.01592306s" podCreationTimestamp="2025-10-07 23:28:08 +0000 UTC" firstStartedPulling="2025-10-07 23:28:09.95334384 +0000 UTC m=+4763.756041913" lastFinishedPulling="2025-10-07 23:28:12.380057275 +0000 UTC m=+4766.182755348" observedRunningTime="2025-10-07 23:28:13.010224649 +0000 UTC m=+4766.812922752" watchObservedRunningTime="2025-10-07 23:28:13.01592306 +0000 UTC m=+4766.818621143" Oct 07 23:28:18 crc kubenswrapper[4871]: I1007 23:28:18.752553 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:18 crc kubenswrapper[4871]: I1007 23:28:18.753424 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:18 crc kubenswrapper[4871]: I1007 23:28:18.817732 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:19 crc kubenswrapper[4871]: I1007 23:28:19.221422 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:19 crc kubenswrapper[4871]: I1007 23:28:19.292145 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2ds8g"] Oct 07 23:28:21 crc kubenswrapper[4871]: I1007 23:28:21.069048 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2ds8g" podUID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerName="registry-server" containerID="cri-o://3d3dad8bd943cffdec74358f92205f5ef6e3354d6f169696b22c4f86162de8bb" gracePeriod=2 Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.091355 4871 generic.go:334] "Generic (PLEG): container finished" podID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerID="3d3dad8bd943cffdec74358f92205f5ef6e3354d6f169696b22c4f86162de8bb" exitCode=0 Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.091672 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ds8g" event={"ID":"130495b5-b37a-4e34-ad8d-b867ee3838f9","Type":"ContainerDied","Data":"3d3dad8bd943cffdec74358f92205f5ef6e3354d6f169696b22c4f86162de8bb"} Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.091711 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2ds8g" event={"ID":"130495b5-b37a-4e34-ad8d-b867ee3838f9","Type":"ContainerDied","Data":"b6822a9481b6a423dd6665c0a5e0f367f0932819bce04a1379ea62361161236e"} Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.091730 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6822a9481b6a423dd6665c0a5e0f367f0932819bce04a1379ea62361161236e" Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.094316 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.244823 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-utilities\") pod \"130495b5-b37a-4e34-ad8d-b867ee3838f9\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.244951 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-catalog-content\") pod \"130495b5-b37a-4e34-ad8d-b867ee3838f9\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.245008 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjg7k\" (UniqueName: \"kubernetes.io/projected/130495b5-b37a-4e34-ad8d-b867ee3838f9-kube-api-access-qjg7k\") pod \"130495b5-b37a-4e34-ad8d-b867ee3838f9\" (UID: \"130495b5-b37a-4e34-ad8d-b867ee3838f9\") " Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.246732 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-utilities" (OuterVolumeSpecName: "utilities") pod "130495b5-b37a-4e34-ad8d-b867ee3838f9" (UID: "130495b5-b37a-4e34-ad8d-b867ee3838f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.250352 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/130495b5-b37a-4e34-ad8d-b867ee3838f9-kube-api-access-qjg7k" (OuterVolumeSpecName: "kube-api-access-qjg7k") pod "130495b5-b37a-4e34-ad8d-b867ee3838f9" (UID: "130495b5-b37a-4e34-ad8d-b867ee3838f9"). InnerVolumeSpecName "kube-api-access-qjg7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.346585 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.346911 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjg7k\" (UniqueName: \"kubernetes.io/projected/130495b5-b37a-4e34-ad8d-b867ee3838f9-kube-api-access-qjg7k\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.360150 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "130495b5-b37a-4e34-ad8d-b867ee3838f9" (UID: "130495b5-b37a-4e34-ad8d-b867ee3838f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:28:23 crc kubenswrapper[4871]: I1007 23:28:23.448388 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130495b5-b37a-4e34-ad8d-b867ee3838f9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:24 crc kubenswrapper[4871]: I1007 23:28:24.109292 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2ds8g" Oct 07 23:28:24 crc kubenswrapper[4871]: I1007 23:28:24.164532 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2ds8g"] Oct 07 23:28:24 crc kubenswrapper[4871]: I1007 23:28:24.181706 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2ds8g"] Oct 07 23:28:24 crc kubenswrapper[4871]: I1007 23:28:24.998965 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="130495b5-b37a-4e34-ad8d-b867ee3838f9" path="/var/lib/kubelet/pods/130495b5-b37a-4e34-ad8d-b867ee3838f9/volumes" Oct 07 23:28:25 crc kubenswrapper[4871]: I1007 23:28:25.137059 4871 generic.go:334] "Generic (PLEG): container finished" podID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" containerID="648fc38c263a9fdb4358e253bd667edffdaa9b9f99c5658764ca1f46e6ea0666" exitCode=0 Oct 07 23:28:25 crc kubenswrapper[4871]: I1007 23:28:25.137130 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"116a6f25-47bf-4d65-aaf8-d3efe5bc8162","Type":"ContainerDied","Data":"648fc38c263a9fdb4358e253bd667edffdaa9b9f99c5658764ca1f46e6ea0666"} Oct 07 23:28:26 crc kubenswrapper[4871]: I1007 23:28:26.145307 4871 generic.go:334] "Generic (PLEG): container finished" podID="7294a0d0-936c-4501-9afe-9ab306906ad2" containerID="a5288a75ee8a01032fe8be52619292a8b1261515d2c65c655e9b313a4c3e34c3" exitCode=0 Oct 07 23:28:26 crc kubenswrapper[4871]: I1007 23:28:26.145402 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7294a0d0-936c-4501-9afe-9ab306906ad2","Type":"ContainerDied","Data":"a5288a75ee8a01032fe8be52619292a8b1261515d2c65c655e9b313a4c3e34c3"} Oct 07 23:28:26 crc kubenswrapper[4871]: I1007 23:28:26.147568 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"116a6f25-47bf-4d65-aaf8-d3efe5bc8162","Type":"ContainerStarted","Data":"5711e865e208fb093e49b27dd73439e89e53de06957f1d48127cb5d9e0357ba0"} Oct 07 23:28:26 crc kubenswrapper[4871]: I1007 23:28:26.147765 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 07 23:28:26 crc kubenswrapper[4871]: I1007 23:28:26.205275 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.205258419 podStartE2EDuration="38.205258419s" podCreationTimestamp="2025-10-07 23:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:28:26.201833929 +0000 UTC m=+4780.004532012" watchObservedRunningTime="2025-10-07 23:28:26.205258419 +0000 UTC m=+4780.007956492" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.159130 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7294a0d0-936c-4501-9afe-9ab306906ad2","Type":"ContainerStarted","Data":"0c93463ceb81ee6ad5efa72352bcfcf172b81eec903295446b2a3e57b4569122"} Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.160161 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.192821 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.19278503 podStartE2EDuration="39.19278503s" podCreationTimestamp="2025-10-07 23:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:28:27.190536251 +0000 UTC m=+4780.993234364" watchObservedRunningTime="2025-10-07 23:28:27.19278503 +0000 UTC m=+4780.995483113" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.314152 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b5nln"] Oct 07 23:28:27 crc kubenswrapper[4871]: E1007 23:28:27.314597 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerName="extract-utilities" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.314625 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerName="extract-utilities" Oct 07 23:28:27 crc kubenswrapper[4871]: E1007 23:28:27.314653 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerName="extract-content" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.314665 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerName="extract-content" Oct 07 23:28:27 crc kubenswrapper[4871]: E1007 23:28:27.314683 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerName="registry-server" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.314695 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerName="registry-server" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.314956 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="130495b5-b37a-4e34-ad8d-b867ee3838f9" containerName="registry-server" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.316688 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.331155 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5nln"] Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.422235 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-utilities\") pod \"redhat-marketplace-b5nln\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.422337 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9gbr\" (UniqueName: \"kubernetes.io/projected/aeb68bfb-76d5-467a-9162-96377cc07dfb-kube-api-access-n9gbr\") pod \"redhat-marketplace-b5nln\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.422451 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-catalog-content\") pod \"redhat-marketplace-b5nln\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.524163 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-catalog-content\") pod \"redhat-marketplace-b5nln\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.524399 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-utilities\") pod \"redhat-marketplace-b5nln\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.524486 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9gbr\" (UniqueName: \"kubernetes.io/projected/aeb68bfb-76d5-467a-9162-96377cc07dfb-kube-api-access-n9gbr\") pod \"redhat-marketplace-b5nln\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.525068 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-utilities\") pod \"redhat-marketplace-b5nln\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.525468 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-catalog-content\") pod \"redhat-marketplace-b5nln\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.545237 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9gbr\" (UniqueName: \"kubernetes.io/projected/aeb68bfb-76d5-467a-9162-96377cc07dfb-kube-api-access-n9gbr\") pod \"redhat-marketplace-b5nln\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:27 crc kubenswrapper[4871]: I1007 23:28:27.674885 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:28 crc kubenswrapper[4871]: I1007 23:28:28.139083 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5nln"] Oct 07 23:28:28 crc kubenswrapper[4871]: I1007 23:28:28.167592 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5nln" event={"ID":"aeb68bfb-76d5-467a-9162-96377cc07dfb","Type":"ContainerStarted","Data":"4f14fb5b914a8d54b03d47045cb658a16efb673fe7b1e39c09a553e75fb61a6a"} Oct 07 23:28:29 crc kubenswrapper[4871]: I1007 23:28:29.177710 4871 generic.go:334] "Generic (PLEG): container finished" podID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerID="66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7" exitCode=0 Oct 07 23:28:29 crc kubenswrapper[4871]: I1007 23:28:29.177866 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5nln" event={"ID":"aeb68bfb-76d5-467a-9162-96377cc07dfb","Type":"ContainerDied","Data":"66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7"} Oct 07 23:28:30 crc kubenswrapper[4871]: I1007 23:28:30.187957 4871 generic.go:334] "Generic (PLEG): container finished" podID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerID="28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193" exitCode=0 Oct 07 23:28:30 crc kubenswrapper[4871]: I1007 23:28:30.188051 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5nln" event={"ID":"aeb68bfb-76d5-467a-9162-96377cc07dfb","Type":"ContainerDied","Data":"28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193"} Oct 07 23:28:31 crc kubenswrapper[4871]: I1007 23:28:31.201723 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5nln" event={"ID":"aeb68bfb-76d5-467a-9162-96377cc07dfb","Type":"ContainerStarted","Data":"da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a"} Oct 07 23:28:31 crc kubenswrapper[4871]: I1007 23:28:31.232011 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b5nln" podStartSLOduration=2.751403185 podStartE2EDuration="4.231982163s" podCreationTimestamp="2025-10-07 23:28:27 +0000 UTC" firstStartedPulling="2025-10-07 23:28:29.179823467 +0000 UTC m=+4782.982521580" lastFinishedPulling="2025-10-07 23:28:30.660402485 +0000 UTC m=+4784.463100558" observedRunningTime="2025-10-07 23:28:31.225529972 +0000 UTC m=+4785.028228045" watchObservedRunningTime="2025-10-07 23:28:31.231982163 +0000 UTC m=+4785.034680276" Oct 07 23:28:35 crc kubenswrapper[4871]: I1007 23:28:35.511907 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:28:35 crc kubenswrapper[4871]: I1007 23:28:35.512243 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:28:37 crc kubenswrapper[4871]: I1007 23:28:37.675977 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:37 crc kubenswrapper[4871]: I1007 23:28:37.676438 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:37 crc kubenswrapper[4871]: I1007 23:28:37.753493 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:38 crc kubenswrapper[4871]: I1007 23:28:38.351502 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:38 crc kubenswrapper[4871]: I1007 23:28:38.405485 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5nln"] Oct 07 23:28:39 crc kubenswrapper[4871]: I1007 23:28:39.969659 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.153066 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.285510 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b5nln" podUID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerName="registry-server" containerID="cri-o://da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a" gracePeriod=2 Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.747413 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.838276 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9gbr\" (UniqueName: \"kubernetes.io/projected/aeb68bfb-76d5-467a-9162-96377cc07dfb-kube-api-access-n9gbr\") pod \"aeb68bfb-76d5-467a-9162-96377cc07dfb\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.838398 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-catalog-content\") pod \"aeb68bfb-76d5-467a-9162-96377cc07dfb\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.838419 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-utilities\") pod \"aeb68bfb-76d5-467a-9162-96377cc07dfb\" (UID: \"aeb68bfb-76d5-467a-9162-96377cc07dfb\") " Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.839497 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-utilities" (OuterVolumeSpecName: "utilities") pod "aeb68bfb-76d5-467a-9162-96377cc07dfb" (UID: "aeb68bfb-76d5-467a-9162-96377cc07dfb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.845206 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb68bfb-76d5-467a-9162-96377cc07dfb-kube-api-access-n9gbr" (OuterVolumeSpecName: "kube-api-access-n9gbr") pod "aeb68bfb-76d5-467a-9162-96377cc07dfb" (UID: "aeb68bfb-76d5-467a-9162-96377cc07dfb"). InnerVolumeSpecName "kube-api-access-n9gbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.872768 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aeb68bfb-76d5-467a-9162-96377cc07dfb" (UID: "aeb68bfb-76d5-467a-9162-96377cc07dfb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.940102 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9gbr\" (UniqueName: \"kubernetes.io/projected/aeb68bfb-76d5-467a-9162-96377cc07dfb-kube-api-access-n9gbr\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.940137 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:40 crc kubenswrapper[4871]: I1007 23:28:40.940146 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb68bfb-76d5-467a-9162-96377cc07dfb-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.294057 4871 generic.go:334] "Generic (PLEG): container finished" podID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerID="da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a" exitCode=0 Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.294105 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5nln" event={"ID":"aeb68bfb-76d5-467a-9162-96377cc07dfb","Type":"ContainerDied","Data":"da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a"} Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.294130 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5nln" event={"ID":"aeb68bfb-76d5-467a-9162-96377cc07dfb","Type":"ContainerDied","Data":"4f14fb5b914a8d54b03d47045cb658a16efb673fe7b1e39c09a553e75fb61a6a"} Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.294145 4871 scope.go:117] "RemoveContainer" containerID="da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.294254 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5nln" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.319326 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5nln"] Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.319423 4871 scope.go:117] "RemoveContainer" containerID="28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.323136 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5nln"] Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.349745 4871 scope.go:117] "RemoveContainer" containerID="66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.367340 4871 scope.go:117] "RemoveContainer" containerID="da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a" Oct 07 23:28:41 crc kubenswrapper[4871]: E1007 23:28:41.367752 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a\": container with ID starting with da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a not found: ID does not exist" containerID="da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.367784 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a"} err="failed to get container status \"da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a\": rpc error: code = NotFound desc = could not find container \"da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a\": container with ID starting with da90c91be14d9abbae3b9a31f5db7544fad072282a8e428592a5f7080bce5b5a not found: ID does not exist" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.367829 4871 scope.go:117] "RemoveContainer" containerID="28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193" Oct 07 23:28:41 crc kubenswrapper[4871]: E1007 23:28:41.368029 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193\": container with ID starting with 28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193 not found: ID does not exist" containerID="28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.368051 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193"} err="failed to get container status \"28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193\": rpc error: code = NotFound desc = could not find container \"28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193\": container with ID starting with 28bd2f571618365644beaf90aea59c45f1a9d9dd5fac6163c79c1478e8eef193 not found: ID does not exist" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.368068 4871 scope.go:117] "RemoveContainer" containerID="66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7" Oct 07 23:28:41 crc kubenswrapper[4871]: E1007 23:28:41.368276 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7\": container with ID starting with 66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7 not found: ID does not exist" containerID="66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7" Oct 07 23:28:41 crc kubenswrapper[4871]: I1007 23:28:41.368299 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7"} err="failed to get container status \"66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7\": rpc error: code = NotFound desc = could not find container \"66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7\": container with ID starting with 66de0eb0aa476bc282decd6005d413f873ee4d8eaf5506d7a0d21e6e38c048a7 not found: ID does not exist" Oct 07 23:28:42 crc kubenswrapper[4871]: I1007 23:28:42.996924 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeb68bfb-76d5-467a-9162-96377cc07dfb" path="/var/lib/kubelet/pods/aeb68bfb-76d5-467a-9162-96377cc07dfb/volumes" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.486614 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-lbns8"] Oct 07 23:28:45 crc kubenswrapper[4871]: E1007 23:28:45.487263 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerName="registry-server" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.487278 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerName="registry-server" Oct 07 23:28:45 crc kubenswrapper[4871]: E1007 23:28:45.487296 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerName="extract-utilities" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.487304 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerName="extract-utilities" Oct 07 23:28:45 crc kubenswrapper[4871]: E1007 23:28:45.487320 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerName="extract-content" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.487329 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerName="extract-content" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.487524 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb68bfb-76d5-467a-9162-96377cc07dfb" containerName="registry-server" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.488424 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.502534 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-lbns8"] Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.616689 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-config\") pod \"dnsmasq-dns-5fdc957c47-lbns8\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.616764 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-dns-svc\") pod \"dnsmasq-dns-5fdc957c47-lbns8\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.617019 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwmv8\" (UniqueName: \"kubernetes.io/projected/5532d18c-3b32-4d80-9db3-1e20e40f4c34-kube-api-access-kwmv8\") pod \"dnsmasq-dns-5fdc957c47-lbns8\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.718826 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwmv8\" (UniqueName: \"kubernetes.io/projected/5532d18c-3b32-4d80-9db3-1e20e40f4c34-kube-api-access-kwmv8\") pod \"dnsmasq-dns-5fdc957c47-lbns8\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.719028 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-config\") pod \"dnsmasq-dns-5fdc957c47-lbns8\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.719075 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-dns-svc\") pod \"dnsmasq-dns-5fdc957c47-lbns8\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.720191 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-config\") pod \"dnsmasq-dns-5fdc957c47-lbns8\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.720308 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-dns-svc\") pod \"dnsmasq-dns-5fdc957c47-lbns8\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.746340 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwmv8\" (UniqueName: \"kubernetes.io/projected/5532d18c-3b32-4d80-9db3-1e20e40f4c34-kube-api-access-kwmv8\") pod \"dnsmasq-dns-5fdc957c47-lbns8\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:45 crc kubenswrapper[4871]: I1007 23:28:45.820689 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:46 crc kubenswrapper[4871]: I1007 23:28:46.049388 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-lbns8"] Oct 07 23:28:46 crc kubenswrapper[4871]: I1007 23:28:46.322580 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 23:28:46 crc kubenswrapper[4871]: I1007 23:28:46.346775 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" event={"ID":"5532d18c-3b32-4d80-9db3-1e20e40f4c34","Type":"ContainerStarted","Data":"6a556f5d5b3270783ac818fb5c7b1a6d78feff8e1821398c06ad2d7487524563"} Oct 07 23:28:47 crc kubenswrapper[4871]: I1007 23:28:47.106430 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 23:28:47 crc kubenswrapper[4871]: I1007 23:28:47.355024 4871 generic.go:334] "Generic (PLEG): container finished" podID="5532d18c-3b32-4d80-9db3-1e20e40f4c34" containerID="f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8" exitCode=0 Oct 07 23:28:47 crc kubenswrapper[4871]: I1007 23:28:47.355075 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" event={"ID":"5532d18c-3b32-4d80-9db3-1e20e40f4c34","Type":"ContainerDied","Data":"f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8"} Oct 07 23:28:48 crc kubenswrapper[4871]: I1007 23:28:48.082927 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" containerName="rabbitmq" containerID="cri-o://5711e865e208fb093e49b27dd73439e89e53de06957f1d48127cb5d9e0357ba0" gracePeriod=604799 Oct 07 23:28:48 crc kubenswrapper[4871]: I1007 23:28:48.366197 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" event={"ID":"5532d18c-3b32-4d80-9db3-1e20e40f4c34","Type":"ContainerStarted","Data":"299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9"} Oct 07 23:28:48 crc kubenswrapper[4871]: I1007 23:28:48.367055 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:48 crc kubenswrapper[4871]: I1007 23:28:48.386535 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" podStartSLOduration=3.386512507 podStartE2EDuration="3.386512507s" podCreationTimestamp="2025-10-07 23:28:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:28:48.384203116 +0000 UTC m=+4802.186901209" watchObservedRunningTime="2025-10-07 23:28:48.386512507 +0000 UTC m=+4802.189210590" Oct 07 23:28:48 crc kubenswrapper[4871]: I1007 23:28:48.913410 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="7294a0d0-936c-4501-9afe-9ab306906ad2" containerName="rabbitmq" containerID="cri-o://0c93463ceb81ee6ad5efa72352bcfcf172b81eec903295446b2a3e57b4569122" gracePeriod=604799 Oct 07 23:28:49 crc kubenswrapper[4871]: I1007 23:28:49.965248 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.234:5672: connect: connection refused" Oct 07 23:28:50 crc kubenswrapper[4871]: I1007 23:28:50.149729 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="7294a0d0-936c-4501-9afe-9ab306906ad2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.235:5672: connect: connection refused" Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.438480 4871 generic.go:334] "Generic (PLEG): container finished" podID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" containerID="5711e865e208fb093e49b27dd73439e89e53de06957f1d48127cb5d9e0357ba0" exitCode=0 Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.438584 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"116a6f25-47bf-4d65-aaf8-d3efe5bc8162","Type":"ContainerDied","Data":"5711e865e208fb093e49b27dd73439e89e53de06957f1d48127cb5d9e0357ba0"} Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.823351 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.986932 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") pod \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.987275 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-plugins\") pod \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.987342 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-erlang-cookie\") pod \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.987373 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-erlang-cookie-secret\") pod \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.987395 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-plugins-conf\") pod \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.987422 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-server-conf\") pod \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.987531 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-pod-info\") pod \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.987561 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-confd\") pod \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.987590 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9ft8\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-kube-api-access-t9ft8\") pod \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\" (UID: \"116a6f25-47bf-4d65-aaf8-d3efe5bc8162\") " Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.988207 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "116a6f25-47bf-4d65-aaf8-d3efe5bc8162" (UID: "116a6f25-47bf-4d65-aaf8-d3efe5bc8162"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.988759 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "116a6f25-47bf-4d65-aaf8-d3efe5bc8162" (UID: "116a6f25-47bf-4d65-aaf8-d3efe5bc8162"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.989263 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "116a6f25-47bf-4d65-aaf8-d3efe5bc8162" (UID: "116a6f25-47bf-4d65-aaf8-d3efe5bc8162"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.995549 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "116a6f25-47bf-4d65-aaf8-d3efe5bc8162" (UID: "116a6f25-47bf-4d65-aaf8-d3efe5bc8162"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:28:54 crc kubenswrapper[4871]: I1007 23:28:54.998086 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-kube-api-access-t9ft8" (OuterVolumeSpecName: "kube-api-access-t9ft8") pod "116a6f25-47bf-4d65-aaf8-d3efe5bc8162" (UID: "116a6f25-47bf-4d65-aaf8-d3efe5bc8162"). InnerVolumeSpecName "kube-api-access-t9ft8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.004368 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-pod-info" (OuterVolumeSpecName: "pod-info") pod "116a6f25-47bf-4d65-aaf8-d3efe5bc8162" (UID: "116a6f25-47bf-4d65-aaf8-d3efe5bc8162"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.015019 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1" (OuterVolumeSpecName: "persistence") pod "116a6f25-47bf-4d65-aaf8-d3efe5bc8162" (UID: "116a6f25-47bf-4d65-aaf8-d3efe5bc8162"). InnerVolumeSpecName "pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.020553 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-server-conf" (OuterVolumeSpecName: "server-conf") pod "116a6f25-47bf-4d65-aaf8-d3efe5bc8162" (UID: "116a6f25-47bf-4d65-aaf8-d3efe5bc8162"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.089904 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.089956 4871 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.089974 4871 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.089994 4871 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-server-conf\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.090159 4871 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-pod-info\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.090193 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9ft8\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-kube-api-access-t9ft8\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.090270 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") on node \"crc\" " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.090290 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.111720 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "116a6f25-47bf-4d65-aaf8-d3efe5bc8162" (UID: "116a6f25-47bf-4d65-aaf8-d3efe5bc8162"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.116786 4871 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.117014 4871 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1") on node "crc" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.191609 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/116a6f25-47bf-4d65-aaf8-d3efe5bc8162-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.191638 4871 reconciler_common.go:293] "Volume detached for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.455455 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"116a6f25-47bf-4d65-aaf8-d3efe5bc8162","Type":"ContainerDied","Data":"e7baa8626363e1db9c16fee18e8b24524795e24fc21eeed12a71e26782d2ea5a"} Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.455518 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.455625 4871 scope.go:117] "RemoveContainer" containerID="5711e865e208fb093e49b27dd73439e89e53de06957f1d48127cb5d9e0357ba0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.462085 4871 generic.go:334] "Generic (PLEG): container finished" podID="7294a0d0-936c-4501-9afe-9ab306906ad2" containerID="0c93463ceb81ee6ad5efa72352bcfcf172b81eec903295446b2a3e57b4569122" exitCode=0 Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.462195 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7294a0d0-936c-4501-9afe-9ab306906ad2","Type":"ContainerDied","Data":"0c93463ceb81ee6ad5efa72352bcfcf172b81eec903295446b2a3e57b4569122"} Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.462536 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7294a0d0-936c-4501-9afe-9ab306906ad2","Type":"ContainerDied","Data":"22e13d38d8572dfbb80aa459e04eeea4b464ef8bc97202891e45f0382459a3b7"} Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.462562 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22e13d38d8572dfbb80aa459e04eeea4b464ef8bc97202891e45f0382459a3b7" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.516495 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.527658 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.527810 4871 scope.go:117] "RemoveContainer" containerID="648fc38c263a9fdb4358e253bd667edffdaa9b9f99c5658764ca1f46e6ea0666" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.532923 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.574782 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 23:28:55 crc kubenswrapper[4871]: E1007 23:28:55.575476 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7294a0d0-936c-4501-9afe-9ab306906ad2" containerName="setup-container" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.575512 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7294a0d0-936c-4501-9afe-9ab306906ad2" containerName="setup-container" Oct 07 23:28:55 crc kubenswrapper[4871]: E1007 23:28:55.575552 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7294a0d0-936c-4501-9afe-9ab306906ad2" containerName="rabbitmq" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.575566 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7294a0d0-936c-4501-9afe-9ab306906ad2" containerName="rabbitmq" Oct 07 23:28:55 crc kubenswrapper[4871]: E1007 23:28:55.575593 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" containerName="rabbitmq" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.575606 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" containerName="rabbitmq" Oct 07 23:28:55 crc kubenswrapper[4871]: E1007 23:28:55.575651 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" containerName="setup-container" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.575666 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" containerName="setup-container" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.576026 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7294a0d0-936c-4501-9afe-9ab306906ad2" containerName="rabbitmq" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.576050 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" containerName="rabbitmq" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.577594 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.591095 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.591297 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.591430 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.591541 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.591644 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-qk4qd" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.599905 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-plugins\") pod \"7294a0d0-936c-4501-9afe-9ab306906ad2\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.600126 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7294a0d0-936c-4501-9afe-9ab306906ad2-erlang-cookie-secret\") pod \"7294a0d0-936c-4501-9afe-9ab306906ad2\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.600176 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-confd\") pod \"7294a0d0-936c-4501-9afe-9ab306906ad2\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.600215 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-plugins-conf\") pod \"7294a0d0-936c-4501-9afe-9ab306906ad2\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.600263 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-erlang-cookie\") pod \"7294a0d0-936c-4501-9afe-9ab306906ad2\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.600302 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-server-conf\") pod \"7294a0d0-936c-4501-9afe-9ab306906ad2\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.600346 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52nk6\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-kube-api-access-52nk6\") pod \"7294a0d0-936c-4501-9afe-9ab306906ad2\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.600399 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7294a0d0-936c-4501-9afe-9ab306906ad2-pod-info\") pod \"7294a0d0-936c-4501-9afe-9ab306906ad2\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.600623 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") pod \"7294a0d0-936c-4501-9afe-9ab306906ad2\" (UID: \"7294a0d0-936c-4501-9afe-9ab306906ad2\") " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.602591 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7294a0d0-936c-4501-9afe-9ab306906ad2" (UID: "7294a0d0-936c-4501-9afe-9ab306906ad2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.602903 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7294a0d0-936c-4501-9afe-9ab306906ad2" (UID: "7294a0d0-936c-4501-9afe-9ab306906ad2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.604869 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.606049 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7294a0d0-936c-4501-9afe-9ab306906ad2" (UID: "7294a0d0-936c-4501-9afe-9ab306906ad2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.609113 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-kube-api-access-52nk6" (OuterVolumeSpecName: "kube-api-access-52nk6") pod "7294a0d0-936c-4501-9afe-9ab306906ad2" (UID: "7294a0d0-936c-4501-9afe-9ab306906ad2"). InnerVolumeSpecName "kube-api-access-52nk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.609476 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7294a0d0-936c-4501-9afe-9ab306906ad2-pod-info" (OuterVolumeSpecName: "pod-info") pod "7294a0d0-936c-4501-9afe-9ab306906ad2" (UID: "7294a0d0-936c-4501-9afe-9ab306906ad2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.614114 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7294a0d0-936c-4501-9afe-9ab306906ad2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7294a0d0-936c-4501-9afe-9ab306906ad2" (UID: "7294a0d0-936c-4501-9afe-9ab306906ad2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.618236 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373" (OuterVolumeSpecName: "persistence") pod "7294a0d0-936c-4501-9afe-9ab306906ad2" (UID: "7294a0d0-936c-4501-9afe-9ab306906ad2"). InnerVolumeSpecName "pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.638491 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-server-conf" (OuterVolumeSpecName: "server-conf") pod "7294a0d0-936c-4501-9afe-9ab306906ad2" (UID: "7294a0d0-936c-4501-9afe-9ab306906ad2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.695429 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7294a0d0-936c-4501-9afe-9ab306906ad2" (UID: "7294a0d0-936c-4501-9afe-9ab306906ad2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.702259 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.702318 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.702371 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.702400 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.702627 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.702850 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.702931 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg2pz\" (UniqueName: \"kubernetes.io/projected/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-kube-api-access-cg2pz\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703035 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703088 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703216 4871 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7294a0d0-936c-4501-9afe-9ab306906ad2-pod-info\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703257 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") on node \"crc\" " Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703281 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703304 4871 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7294a0d0-936c-4501-9afe-9ab306906ad2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703321 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703337 4871 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703355 4871 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7294a0d0-936c-4501-9afe-9ab306906ad2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703374 4871 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7294a0d0-936c-4501-9afe-9ab306906ad2-server-conf\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.703392 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52nk6\" (UniqueName: \"kubernetes.io/projected/7294a0d0-936c-4501-9afe-9ab306906ad2-kube-api-access-52nk6\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.722686 4871 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.722976 4871 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373") on node "crc" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.805113 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.805427 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.805512 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.805606 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.805719 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.805825 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg2pz\" (UniqueName: \"kubernetes.io/projected/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-kube-api-access-cg2pz\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.805925 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.805995 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.806074 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.806180 4871 reconciler_common.go:293] "Volume detached for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.806491 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.806756 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.806878 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.806881 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.808902 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.810185 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.810304 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.810340 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2c8faa66fce466b48b425a0e4cbd921197e1b803b01adab0d2859d6a441f6bef/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.811422 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.821724 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg2pz\" (UniqueName: \"kubernetes.io/projected/2b9bb1f5-e47a-48af-9a8b-4b766881f56f-kube-api-access-cg2pz\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.821945 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.862977 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-977f9fb8-c1bf-4eaa-804b-c6f1058de4d1\") pod \"rabbitmq-server-0\" (UID: \"2b9bb1f5-e47a-48af-9a8b-4b766881f56f\") " pod="openstack/rabbitmq-server-0" Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.879669 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-j7kwj"] Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.881736 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" podUID="e19bcc51-e926-424e-aecf-b6985e55462e" containerName="dnsmasq-dns" containerID="cri-o://d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f" gracePeriod=10 Oct 07 23:28:55 crc kubenswrapper[4871]: I1007 23:28:55.959481 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.261127 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.380248 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.419014 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-config\") pod \"e19bcc51-e926-424e-aecf-b6985e55462e\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.419084 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbh9m\" (UniqueName: \"kubernetes.io/projected/e19bcc51-e926-424e-aecf-b6985e55462e-kube-api-access-sbh9m\") pod \"e19bcc51-e926-424e-aecf-b6985e55462e\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.419209 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-dns-svc\") pod \"e19bcc51-e926-424e-aecf-b6985e55462e\" (UID: \"e19bcc51-e926-424e-aecf-b6985e55462e\") " Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.424058 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e19bcc51-e926-424e-aecf-b6985e55462e-kube-api-access-sbh9m" (OuterVolumeSpecName: "kube-api-access-sbh9m") pod "e19bcc51-e926-424e-aecf-b6985e55462e" (UID: "e19bcc51-e926-424e-aecf-b6985e55462e"). InnerVolumeSpecName "kube-api-access-sbh9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.462756 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-config" (OuterVolumeSpecName: "config") pod "e19bcc51-e926-424e-aecf-b6985e55462e" (UID: "e19bcc51-e926-424e-aecf-b6985e55462e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.470111 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b9bb1f5-e47a-48af-9a8b-4b766881f56f","Type":"ContainerStarted","Data":"b1d40ceba0869e9abb03cee089bafdf307e97901ecaa2ec5d4006544be07257b"} Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.472431 4871 generic.go:334] "Generic (PLEG): container finished" podID="e19bcc51-e926-424e-aecf-b6985e55462e" containerID="d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f" exitCode=0 Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.472507 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" event={"ID":"e19bcc51-e926-424e-aecf-b6985e55462e","Type":"ContainerDied","Data":"d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f"} Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.472557 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.472582 4871 scope.go:117] "RemoveContainer" containerID="d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.472573 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.472565 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-j7kwj" event={"ID":"e19bcc51-e926-424e-aecf-b6985e55462e","Type":"ContainerDied","Data":"77e49fa68190e6b5ef5ff4b46270deaa9c939489989f7b2a3bee07ab6affabf7"} Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.478131 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e19bcc51-e926-424e-aecf-b6985e55462e" (UID: "e19bcc51-e926-424e-aecf-b6985e55462e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.512161 4871 scope.go:117] "RemoveContainer" containerID="37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.523348 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.523380 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19bcc51-e926-424e-aecf-b6985e55462e-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.523393 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbh9m\" (UniqueName: \"kubernetes.io/projected/e19bcc51-e926-424e-aecf-b6985e55462e-kube-api-access-sbh9m\") on node \"crc\" DevicePath \"\"" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.532116 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.541531 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.551955 4871 scope.go:117] "RemoveContainer" containerID="d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f" Oct 07 23:28:56 crc kubenswrapper[4871]: E1007 23:28:56.552701 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f\": container with ID starting with d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f not found: ID does not exist" containerID="d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.552725 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f"} err="failed to get container status \"d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f\": rpc error: code = NotFound desc = could not find container \"d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f\": container with ID starting with d5daec1bdaf278d71c7a7d3574024192c309419ca7c75c736a38e8056c9d583f not found: ID does not exist" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.552744 4871 scope.go:117] "RemoveContainer" containerID="37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b" Oct 07 23:28:56 crc kubenswrapper[4871]: E1007 23:28:56.555303 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b\": container with ID starting with 37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b not found: ID does not exist" containerID="37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.555329 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b"} err="failed to get container status \"37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b\": rpc error: code = NotFound desc = could not find container \"37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b\": container with ID starting with 37bc284afb9a7decc76ac03b8d12c60c2bdee42c2b317f7fde1e31c000cfea3b not found: ID does not exist" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.555824 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 23:28:56 crc kubenswrapper[4871]: E1007 23:28:56.556105 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e19bcc51-e926-424e-aecf-b6985e55462e" containerName="init" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.556118 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e19bcc51-e926-424e-aecf-b6985e55462e" containerName="init" Oct 07 23:28:56 crc kubenswrapper[4871]: E1007 23:28:56.556137 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e19bcc51-e926-424e-aecf-b6985e55462e" containerName="dnsmasq-dns" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.556143 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e19bcc51-e926-424e-aecf-b6985e55462e" containerName="dnsmasq-dns" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.556280 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e19bcc51-e926-424e-aecf-b6985e55462e" containerName="dnsmasq-dns" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.557252 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.560269 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.560355 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.560592 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.560753 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-w52c2" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.561017 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.580825 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.624346 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jfxv\" (UniqueName: \"kubernetes.io/projected/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-kube-api-access-9jfxv\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.624388 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.624428 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.624558 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.624596 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.624678 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.624704 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.624727 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.624780 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.726909 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.727029 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.727062 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.727133 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.727169 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.727218 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.727280 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.727393 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jfxv\" (UniqueName: \"kubernetes.io/projected/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-kube-api-access-9jfxv\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.727430 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.728067 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.728165 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.729347 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.730076 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.730682 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.730736 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a649604412d27923c0fe096e2ba7c6ce60c40b8e70d73cbc9cac8dd896ba0a3c/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.733229 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.734469 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.735984 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.747746 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jfxv\" (UniqueName: \"kubernetes.io/projected/50bfcc25-71f9-46ee-b7ab-7be7d920c19f-kube-api-access-9jfxv\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.782296 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-95b1bcd0-9a0c-4dbf-a899-2de5c34b9373\") pod \"rabbitmq-cell1-server-0\" (UID: \"50bfcc25-71f9-46ee-b7ab-7be7d920c19f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.827446 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-j7kwj"] Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.837361 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-j7kwj"] Oct 07 23:28:56 crc kubenswrapper[4871]: I1007 23:28:56.878148 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:28:57 crc kubenswrapper[4871]: I1007 23:28:57.000975 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="116a6f25-47bf-4d65-aaf8-d3efe5bc8162" path="/var/lib/kubelet/pods/116a6f25-47bf-4d65-aaf8-d3efe5bc8162/volumes" Oct 07 23:28:57 crc kubenswrapper[4871]: I1007 23:28:57.003950 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7294a0d0-936c-4501-9afe-9ab306906ad2" path="/var/lib/kubelet/pods/7294a0d0-936c-4501-9afe-9ab306906ad2/volumes" Oct 07 23:28:57 crc kubenswrapper[4871]: I1007 23:28:57.006123 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e19bcc51-e926-424e-aecf-b6985e55462e" path="/var/lib/kubelet/pods/e19bcc51-e926-424e-aecf-b6985e55462e/volumes" Oct 07 23:28:57 crc kubenswrapper[4871]: I1007 23:28:57.959434 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 23:28:57 crc kubenswrapper[4871]: W1007 23:28:57.963933 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50bfcc25_71f9_46ee_b7ab_7be7d920c19f.slice/crio-55b3d1a3094f845274aa47be6e465dddecf7845fea0bfaa933bf3b9a7dce55ee WatchSource:0}: Error finding container 55b3d1a3094f845274aa47be6e465dddecf7845fea0bfaa933bf3b9a7dce55ee: Status 404 returned error can't find the container with id 55b3d1a3094f845274aa47be6e465dddecf7845fea0bfaa933bf3b9a7dce55ee Oct 07 23:28:58 crc kubenswrapper[4871]: I1007 23:28:58.494749 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"50bfcc25-71f9-46ee-b7ab-7be7d920c19f","Type":"ContainerStarted","Data":"55b3d1a3094f845274aa47be6e465dddecf7845fea0bfaa933bf3b9a7dce55ee"} Oct 07 23:28:59 crc kubenswrapper[4871]: I1007 23:28:59.509283 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b9bb1f5-e47a-48af-9a8b-4b766881f56f","Type":"ContainerStarted","Data":"3f3bbb04a40d71cfd2370871df4f97262b34eada7c879ac6052db747ced12ce1"} Oct 07 23:29:00 crc kubenswrapper[4871]: I1007 23:29:00.523567 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"50bfcc25-71f9-46ee-b7ab-7be7d920c19f","Type":"ContainerStarted","Data":"61cd20454b7f835df5f31d6acd7a310076c959d95ec39b7a91300bf8c5a3288d"} Oct 07 23:29:05 crc kubenswrapper[4871]: I1007 23:29:05.513089 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:29:05 crc kubenswrapper[4871]: I1007 23:29:05.513686 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:29:05 crc kubenswrapper[4871]: I1007 23:29:05.513738 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:29:05 crc kubenswrapper[4871]: I1007 23:29:05.514451 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ab5d945d29b218b1fcfe5ad517fbc83535bc60079d91b4857bc595f89a8605f"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:29:05 crc kubenswrapper[4871]: I1007 23:29:05.514522 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://3ab5d945d29b218b1fcfe5ad517fbc83535bc60079d91b4857bc595f89a8605f" gracePeriod=600 Oct 07 23:29:06 crc kubenswrapper[4871]: I1007 23:29:06.594045 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="3ab5d945d29b218b1fcfe5ad517fbc83535bc60079d91b4857bc595f89a8605f" exitCode=0 Oct 07 23:29:06 crc kubenswrapper[4871]: I1007 23:29:06.594126 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"3ab5d945d29b218b1fcfe5ad517fbc83535bc60079d91b4857bc595f89a8605f"} Oct 07 23:29:06 crc kubenswrapper[4871]: I1007 23:29:06.594550 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a"} Oct 07 23:29:06 crc kubenswrapper[4871]: I1007 23:29:06.594582 4871 scope.go:117] "RemoveContainer" containerID="9ff5b37701d9899d46560f0bf3963ba5c852bdd9c2bd7afe478ffc207c72660e" Oct 07 23:29:32 crc kubenswrapper[4871]: I1007 23:29:32.881620 4871 generic.go:334] "Generic (PLEG): container finished" podID="2b9bb1f5-e47a-48af-9a8b-4b766881f56f" containerID="3f3bbb04a40d71cfd2370871df4f97262b34eada7c879ac6052db747ced12ce1" exitCode=0 Oct 07 23:29:32 crc kubenswrapper[4871]: I1007 23:29:32.881745 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b9bb1f5-e47a-48af-9a8b-4b766881f56f","Type":"ContainerDied","Data":"3f3bbb04a40d71cfd2370871df4f97262b34eada7c879ac6052db747ced12ce1"} Oct 07 23:29:33 crc kubenswrapper[4871]: I1007 23:29:33.898161 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b9bb1f5-e47a-48af-9a8b-4b766881f56f","Type":"ContainerStarted","Data":"0093da24412a4754d848764d292f82d45f0e9357d0fec1a05753c694dac65912"} Oct 07 23:29:33 crc kubenswrapper[4871]: I1007 23:29:33.898845 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 07 23:29:33 crc kubenswrapper[4871]: I1007 23:29:33.900515 4871 generic.go:334] "Generic (PLEG): container finished" podID="50bfcc25-71f9-46ee-b7ab-7be7d920c19f" containerID="61cd20454b7f835df5f31d6acd7a310076c959d95ec39b7a91300bf8c5a3288d" exitCode=0 Oct 07 23:29:33 crc kubenswrapper[4871]: I1007 23:29:33.900558 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"50bfcc25-71f9-46ee-b7ab-7be7d920c19f","Type":"ContainerDied","Data":"61cd20454b7f835df5f31d6acd7a310076c959d95ec39b7a91300bf8c5a3288d"} Oct 07 23:29:33 crc kubenswrapper[4871]: I1007 23:29:33.938385 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.938356963 podStartE2EDuration="38.938356963s" podCreationTimestamp="2025-10-07 23:28:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:29:33.923044379 +0000 UTC m=+4847.725742482" watchObservedRunningTime="2025-10-07 23:29:33.938356963 +0000 UTC m=+4847.741055086" Oct 07 23:29:34 crc kubenswrapper[4871]: I1007 23:29:34.909574 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"50bfcc25-71f9-46ee-b7ab-7be7d920c19f","Type":"ContainerStarted","Data":"3284eeea2ba85e905fd72649fe018914e5e6e542eec9bdc7510ce000a4b9e19c"} Oct 07 23:29:34 crc kubenswrapper[4871]: I1007 23:29:34.910947 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:29:34 crc kubenswrapper[4871]: I1007 23:29:34.938024 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.938008094 podStartE2EDuration="38.938008094s" podCreationTimestamp="2025-10-07 23:28:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:29:34.934154462 +0000 UTC m=+4848.736852575" watchObservedRunningTime="2025-10-07 23:29:34.938008094 +0000 UTC m=+4848.740706167" Oct 07 23:29:45 crc kubenswrapper[4871]: I1007 23:29:45.961999 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 07 23:29:46 crc kubenswrapper[4871]: I1007 23:29:46.882186 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 07 23:29:54 crc kubenswrapper[4871]: I1007 23:29:54.613326 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 07 23:29:54 crc kubenswrapper[4871]: I1007 23:29:54.615057 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 07 23:29:54 crc kubenswrapper[4871]: I1007 23:29:54.617438 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rqp8s" Oct 07 23:29:54 crc kubenswrapper[4871]: I1007 23:29:54.629962 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 07 23:29:54 crc kubenswrapper[4871]: I1007 23:29:54.726842 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bht5f\" (UniqueName: \"kubernetes.io/projected/dd1676bd-2019-4786-9589-1e11df6ea746-kube-api-access-bht5f\") pod \"mariadb-client-1-default\" (UID: \"dd1676bd-2019-4786-9589-1e11df6ea746\") " pod="openstack/mariadb-client-1-default" Oct 07 23:29:54 crc kubenswrapper[4871]: I1007 23:29:54.827942 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bht5f\" (UniqueName: \"kubernetes.io/projected/dd1676bd-2019-4786-9589-1e11df6ea746-kube-api-access-bht5f\") pod \"mariadb-client-1-default\" (UID: \"dd1676bd-2019-4786-9589-1e11df6ea746\") " pod="openstack/mariadb-client-1-default" Oct 07 23:29:54 crc kubenswrapper[4871]: I1007 23:29:54.854902 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bht5f\" (UniqueName: \"kubernetes.io/projected/dd1676bd-2019-4786-9589-1e11df6ea746-kube-api-access-bht5f\") pod \"mariadb-client-1-default\" (UID: \"dd1676bd-2019-4786-9589-1e11df6ea746\") " pod="openstack/mariadb-client-1-default" Oct 07 23:29:54 crc kubenswrapper[4871]: I1007 23:29:54.953620 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 07 23:29:56 crc kubenswrapper[4871]: I1007 23:29:55.512325 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 07 23:29:56 crc kubenswrapper[4871]: I1007 23:29:55.532578 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 23:29:56 crc kubenswrapper[4871]: I1007 23:29:56.140126 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"dd1676bd-2019-4786-9589-1e11df6ea746","Type":"ContainerStarted","Data":"fb964cd0ce8b5b84fd64819245315ec68587b2fe68af2714d2b9ff63810e5acc"} Oct 07 23:29:57 crc kubenswrapper[4871]: I1007 23:29:57.153282 4871 generic.go:334] "Generic (PLEG): container finished" podID="dd1676bd-2019-4786-9589-1e11df6ea746" containerID="2f83aa1535305d8ae6af378cb44cb2d64038ff47dc98e1772e581aadba9b3d5d" exitCode=0 Oct 07 23:29:57 crc kubenswrapper[4871]: I1007 23:29:57.153378 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"dd1676bd-2019-4786-9589-1e11df6ea746","Type":"ContainerDied","Data":"2f83aa1535305d8ae6af378cb44cb2d64038ff47dc98e1772e581aadba9b3d5d"} Oct 07 23:29:58 crc kubenswrapper[4871]: I1007 23:29:58.886371 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 07 23:29:58 crc kubenswrapper[4871]: I1007 23:29:58.922963 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_dd1676bd-2019-4786-9589-1e11df6ea746/mariadb-client-1-default/0.log" Oct 07 23:29:58 crc kubenswrapper[4871]: I1007 23:29:58.967114 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 07 23:29:58 crc kubenswrapper[4871]: I1007 23:29:58.972221 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 07 23:29:58 crc kubenswrapper[4871]: I1007 23:29:58.999045 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bht5f\" (UniqueName: \"kubernetes.io/projected/dd1676bd-2019-4786-9589-1e11df6ea746-kube-api-access-bht5f\") pod \"dd1676bd-2019-4786-9589-1e11df6ea746\" (UID: \"dd1676bd-2019-4786-9589-1e11df6ea746\") " Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.010094 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd1676bd-2019-4786-9589-1e11df6ea746-kube-api-access-bht5f" (OuterVolumeSpecName: "kube-api-access-bht5f") pod "dd1676bd-2019-4786-9589-1e11df6ea746" (UID: "dd1676bd-2019-4786-9589-1e11df6ea746"). InnerVolumeSpecName "kube-api-access-bht5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.101248 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bht5f\" (UniqueName: \"kubernetes.io/projected/dd1676bd-2019-4786-9589-1e11df6ea746-kube-api-access-bht5f\") on node \"crc\" DevicePath \"\"" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.181977 4871 scope.go:117] "RemoveContainer" containerID="2f83aa1535305d8ae6af378cb44cb2d64038ff47dc98e1772e581aadba9b3d5d" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.182060 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.432349 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 07 23:29:59 crc kubenswrapper[4871]: E1007 23:29:59.432756 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd1676bd-2019-4786-9589-1e11df6ea746" containerName="mariadb-client-1-default" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.432774 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd1676bd-2019-4786-9589-1e11df6ea746" containerName="mariadb-client-1-default" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.432963 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd1676bd-2019-4786-9589-1e11df6ea746" containerName="mariadb-client-1-default" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.433503 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.436669 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rqp8s" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.446001 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.508704 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7c4q\" (UniqueName: \"kubernetes.io/projected/da923cb6-6af2-4556-86b9-0188d806c9d6-kube-api-access-k7c4q\") pod \"mariadb-client-2-default\" (UID: \"da923cb6-6af2-4556-86b9-0188d806c9d6\") " pod="openstack/mariadb-client-2-default" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.609741 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7c4q\" (UniqueName: \"kubernetes.io/projected/da923cb6-6af2-4556-86b9-0188d806c9d6-kube-api-access-k7c4q\") pod \"mariadb-client-2-default\" (UID: \"da923cb6-6af2-4556-86b9-0188d806c9d6\") " pod="openstack/mariadb-client-2-default" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.640635 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7c4q\" (UniqueName: \"kubernetes.io/projected/da923cb6-6af2-4556-86b9-0188d806c9d6-kube-api-access-k7c4q\") pod \"mariadb-client-2-default\" (UID: \"da923cb6-6af2-4556-86b9-0188d806c9d6\") " pod="openstack/mariadb-client-2-default" Oct 07 23:29:59 crc kubenswrapper[4871]: I1007 23:29:59.768887 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.147286 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk"] Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.149078 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.153049 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.153139 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.153984 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk"] Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.244882 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96b42486-ae7f-4ef5-b2f9-6adbbce86922-config-volume\") pod \"collect-profiles-29331330-rmnfk\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.244934 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96b42486-ae7f-4ef5-b2f9-6adbbce86922-secret-volume\") pod \"collect-profiles-29331330-rmnfk\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.244956 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2vhf\" (UniqueName: \"kubernetes.io/projected/96b42486-ae7f-4ef5-b2f9-6adbbce86922-kube-api-access-c2vhf\") pod \"collect-profiles-29331330-rmnfk\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.337652 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.349756 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96b42486-ae7f-4ef5-b2f9-6adbbce86922-config-volume\") pod \"collect-profiles-29331330-rmnfk\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.349817 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96b42486-ae7f-4ef5-b2f9-6adbbce86922-secret-volume\") pod \"collect-profiles-29331330-rmnfk\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.349851 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2vhf\" (UniqueName: \"kubernetes.io/projected/96b42486-ae7f-4ef5-b2f9-6adbbce86922-kube-api-access-c2vhf\") pod \"collect-profiles-29331330-rmnfk\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.351161 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96b42486-ae7f-4ef5-b2f9-6adbbce86922-config-volume\") pod \"collect-profiles-29331330-rmnfk\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.367855 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96b42486-ae7f-4ef5-b2f9-6adbbce86922-secret-volume\") pod \"collect-profiles-29331330-rmnfk\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.371876 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2vhf\" (UniqueName: \"kubernetes.io/projected/96b42486-ae7f-4ef5-b2f9-6adbbce86922-kube-api-access-c2vhf\") pod \"collect-profiles-29331330-rmnfk\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.471639 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.884341 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk"] Oct 07 23:30:00 crc kubenswrapper[4871]: I1007 23:30:00.992558 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd1676bd-2019-4786-9589-1e11df6ea746" path="/var/lib/kubelet/pods/dd1676bd-2019-4786-9589-1e11df6ea746/volumes" Oct 07 23:30:01 crc kubenswrapper[4871]: I1007 23:30:01.200012 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" event={"ID":"96b42486-ae7f-4ef5-b2f9-6adbbce86922","Type":"ContainerStarted","Data":"6119698f5691f10e9cf19cbcb4553b97df2c01a2fed40ea99c4584dc29736092"} Oct 07 23:30:01 crc kubenswrapper[4871]: I1007 23:30:01.200068 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" event={"ID":"96b42486-ae7f-4ef5-b2f9-6adbbce86922","Type":"ContainerStarted","Data":"73eeaa0109d6af96fd520f6aa7c650761c3e34ace41d3aa2e9667af5d59696de"} Oct 07 23:30:01 crc kubenswrapper[4871]: I1007 23:30:01.202644 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"da923cb6-6af2-4556-86b9-0188d806c9d6","Type":"ContainerStarted","Data":"8718a9f20f2928e2408853b41fe9d6cd12527ccfb70e0b57936a5b6f6cfa37a8"} Oct 07 23:30:01 crc kubenswrapper[4871]: I1007 23:30:01.202671 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"da923cb6-6af2-4556-86b9-0188d806c9d6","Type":"ContainerStarted","Data":"3dd644a6a80c96c95c63a0193822d0ea45cf7cef268153b0a2c812a3d41028a0"} Oct 07 23:30:01 crc kubenswrapper[4871]: I1007 23:30:01.230723 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=2.23069352 podStartE2EDuration="2.23069352s" podCreationTimestamp="2025-10-07 23:29:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:30:01.224513947 +0000 UTC m=+4875.027212050" watchObservedRunningTime="2025-10-07 23:30:01.23069352 +0000 UTC m=+4875.033391633" Oct 07 23:30:02 crc kubenswrapper[4871]: I1007 23:30:02.213884 4871 generic.go:334] "Generic (PLEG): container finished" podID="da923cb6-6af2-4556-86b9-0188d806c9d6" containerID="8718a9f20f2928e2408853b41fe9d6cd12527ccfb70e0b57936a5b6f6cfa37a8" exitCode=0 Oct 07 23:30:02 crc kubenswrapper[4871]: I1007 23:30:02.214342 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"da923cb6-6af2-4556-86b9-0188d806c9d6","Type":"ContainerDied","Data":"8718a9f20f2928e2408853b41fe9d6cd12527ccfb70e0b57936a5b6f6cfa37a8"} Oct 07 23:30:02 crc kubenswrapper[4871]: I1007 23:30:02.216132 4871 generic.go:334] "Generic (PLEG): container finished" podID="96b42486-ae7f-4ef5-b2f9-6adbbce86922" containerID="6119698f5691f10e9cf19cbcb4553b97df2c01a2fed40ea99c4584dc29736092" exitCode=0 Oct 07 23:30:02 crc kubenswrapper[4871]: I1007 23:30:02.216162 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" event={"ID":"96b42486-ae7f-4ef5-b2f9-6adbbce86922","Type":"ContainerDied","Data":"6119698f5691f10e9cf19cbcb4553b97df2c01a2fed40ea99c4584dc29736092"} Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.704749 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.712860 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.763821 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.770375 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.808455 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96b42486-ae7f-4ef5-b2f9-6adbbce86922-config-volume\") pod \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.808708 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2vhf\" (UniqueName: \"kubernetes.io/projected/96b42486-ae7f-4ef5-b2f9-6adbbce86922-kube-api-access-c2vhf\") pod \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.808922 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7c4q\" (UniqueName: \"kubernetes.io/projected/da923cb6-6af2-4556-86b9-0188d806c9d6-kube-api-access-k7c4q\") pod \"da923cb6-6af2-4556-86b9-0188d806c9d6\" (UID: \"da923cb6-6af2-4556-86b9-0188d806c9d6\") " Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.809036 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96b42486-ae7f-4ef5-b2f9-6adbbce86922-secret-volume\") pod \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\" (UID: \"96b42486-ae7f-4ef5-b2f9-6adbbce86922\") " Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.809610 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96b42486-ae7f-4ef5-b2f9-6adbbce86922-config-volume" (OuterVolumeSpecName: "config-volume") pod "96b42486-ae7f-4ef5-b2f9-6adbbce86922" (UID: "96b42486-ae7f-4ef5-b2f9-6adbbce86922"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.815611 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b42486-ae7f-4ef5-b2f9-6adbbce86922-kube-api-access-c2vhf" (OuterVolumeSpecName: "kube-api-access-c2vhf") pod "96b42486-ae7f-4ef5-b2f9-6adbbce86922" (UID: "96b42486-ae7f-4ef5-b2f9-6adbbce86922"). InnerVolumeSpecName "kube-api-access-c2vhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.815985 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da923cb6-6af2-4556-86b9-0188d806c9d6-kube-api-access-k7c4q" (OuterVolumeSpecName: "kube-api-access-k7c4q") pod "da923cb6-6af2-4556-86b9-0188d806c9d6" (UID: "da923cb6-6af2-4556-86b9-0188d806c9d6"). InnerVolumeSpecName "kube-api-access-k7c4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.821594 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b42486-ae7f-4ef5-b2f9-6adbbce86922-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "96b42486-ae7f-4ef5-b2f9-6adbbce86922" (UID: "96b42486-ae7f-4ef5-b2f9-6adbbce86922"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.910905 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96b42486-ae7f-4ef5-b2f9-6adbbce86922-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.911089 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96b42486-ae7f-4ef5-b2f9-6adbbce86922-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.911115 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2vhf\" (UniqueName: \"kubernetes.io/projected/96b42486-ae7f-4ef5-b2f9-6adbbce86922-kube-api-access-c2vhf\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:03 crc kubenswrapper[4871]: I1007 23:30:03.911135 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7c4q\" (UniqueName: \"kubernetes.io/projected/da923cb6-6af2-4556-86b9-0188d806c9d6-kube-api-access-k7c4q\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.241181 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" event={"ID":"96b42486-ae7f-4ef5-b2f9-6adbbce86922","Type":"ContainerDied","Data":"73eeaa0109d6af96fd520f6aa7c650761c3e34ace41d3aa2e9667af5d59696de"} Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.241260 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73eeaa0109d6af96fd520f6aa7c650761c3e34ace41d3aa2e9667af5d59696de" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.241313 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.249345 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dd644a6a80c96c95c63a0193822d0ea45cf7cef268153b0a2c812a3d41028a0" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.249472 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.268830 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 07 23:30:04 crc kubenswrapper[4871]: E1007 23:30:04.269524 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96b42486-ae7f-4ef5-b2f9-6adbbce86922" containerName="collect-profiles" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.269570 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="96b42486-ae7f-4ef5-b2f9-6adbbce86922" containerName="collect-profiles" Oct 07 23:30:04 crc kubenswrapper[4871]: E1007 23:30:04.269621 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da923cb6-6af2-4556-86b9-0188d806c9d6" containerName="mariadb-client-2-default" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.269640 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="da923cb6-6af2-4556-86b9-0188d806c9d6" containerName="mariadb-client-2-default" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.270059 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="da923cb6-6af2-4556-86b9-0188d806c9d6" containerName="mariadb-client-2-default" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.270124 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="96b42486-ae7f-4ef5-b2f9-6adbbce86922" containerName="collect-profiles" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.272105 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.278224 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rqp8s" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.288514 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.318038 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz2mc\" (UniqueName: \"kubernetes.io/projected/7dcdf1a1-102d-4d4c-adf5-1736b0ea1888-kube-api-access-sz2mc\") pod \"mariadb-client-1\" (UID: \"7dcdf1a1-102d-4d4c-adf5-1736b0ea1888\") " pod="openstack/mariadb-client-1" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.419175 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz2mc\" (UniqueName: \"kubernetes.io/projected/7dcdf1a1-102d-4d4c-adf5-1736b0ea1888-kube-api-access-sz2mc\") pod \"mariadb-client-1\" (UID: \"7dcdf1a1-102d-4d4c-adf5-1736b0ea1888\") " pod="openstack/mariadb-client-1" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.435532 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz2mc\" (UniqueName: \"kubernetes.io/projected/7dcdf1a1-102d-4d4c-adf5-1736b0ea1888-kube-api-access-sz2mc\") pod \"mariadb-client-1\" (UID: \"7dcdf1a1-102d-4d4c-adf5-1736b0ea1888\") " pod="openstack/mariadb-client-1" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.665656 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.792213 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94"] Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.796235 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331285-8ss94"] Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.992455 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da923cb6-6af2-4556-86b9-0188d806c9d6" path="/var/lib/kubelet/pods/da923cb6-6af2-4556-86b9-0188d806c9d6/volumes" Oct 07 23:30:04 crc kubenswrapper[4871]: I1007 23:30:04.993663 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8" path="/var/lib/kubelet/pods/eec468c5-17f5-43ef-9fb1-cc93c6c5c3f8/volumes" Oct 07 23:30:05 crc kubenswrapper[4871]: I1007 23:30:05.328497 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 07 23:30:05 crc kubenswrapper[4871]: W1007 23:30:05.341984 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dcdf1a1_102d_4d4c_adf5_1736b0ea1888.slice/crio-0e7543d54477ca57f155ee30eb6a0996ca796b9c3421bde21a4e70bea85f7555 WatchSource:0}: Error finding container 0e7543d54477ca57f155ee30eb6a0996ca796b9c3421bde21a4e70bea85f7555: Status 404 returned error can't find the container with id 0e7543d54477ca57f155ee30eb6a0996ca796b9c3421bde21a4e70bea85f7555 Oct 07 23:30:06 crc kubenswrapper[4871]: I1007 23:30:06.268771 4871 generic.go:334] "Generic (PLEG): container finished" podID="7dcdf1a1-102d-4d4c-adf5-1736b0ea1888" containerID="a857283ba9648d68b58096f8ff8568fa526f8a8ddd8d2fd710436012ceaaef32" exitCode=0 Oct 07 23:30:06 crc kubenswrapper[4871]: I1007 23:30:06.268874 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"7dcdf1a1-102d-4d4c-adf5-1736b0ea1888","Type":"ContainerDied","Data":"a857283ba9648d68b58096f8ff8568fa526f8a8ddd8d2fd710436012ceaaef32"} Oct 07 23:30:06 crc kubenswrapper[4871]: I1007 23:30:06.268915 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"7dcdf1a1-102d-4d4c-adf5-1736b0ea1888","Type":"ContainerStarted","Data":"0e7543d54477ca57f155ee30eb6a0996ca796b9c3421bde21a4e70bea85f7555"} Oct 07 23:30:07 crc kubenswrapper[4871]: I1007 23:30:07.747029 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 07 23:30:07 crc kubenswrapper[4871]: I1007 23:30:07.771665 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_7dcdf1a1-102d-4d4c-adf5-1736b0ea1888/mariadb-client-1/0.log" Oct 07 23:30:07 crc kubenswrapper[4871]: I1007 23:30:07.777905 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz2mc\" (UniqueName: \"kubernetes.io/projected/7dcdf1a1-102d-4d4c-adf5-1736b0ea1888-kube-api-access-sz2mc\") pod \"7dcdf1a1-102d-4d4c-adf5-1736b0ea1888\" (UID: \"7dcdf1a1-102d-4d4c-adf5-1736b0ea1888\") " Oct 07 23:30:07 crc kubenswrapper[4871]: I1007 23:30:07.784486 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dcdf1a1-102d-4d4c-adf5-1736b0ea1888-kube-api-access-sz2mc" (OuterVolumeSpecName: "kube-api-access-sz2mc") pod "7dcdf1a1-102d-4d4c-adf5-1736b0ea1888" (UID: "7dcdf1a1-102d-4d4c-adf5-1736b0ea1888"). InnerVolumeSpecName "kube-api-access-sz2mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:30:07 crc kubenswrapper[4871]: I1007 23:30:07.799305 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 07 23:30:07 crc kubenswrapper[4871]: I1007 23:30:07.814120 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 07 23:30:07 crc kubenswrapper[4871]: I1007 23:30:07.879834 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz2mc\" (UniqueName: \"kubernetes.io/projected/7dcdf1a1-102d-4d4c-adf5-1736b0ea1888-kube-api-access-sz2mc\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.286209 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e7543d54477ca57f155ee30eb6a0996ca796b9c3421bde21a4e70bea85f7555" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.286286 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.340356 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 07 23:30:08 crc kubenswrapper[4871]: E1007 23:30:08.340693 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dcdf1a1-102d-4d4c-adf5-1736b0ea1888" containerName="mariadb-client-1" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.340715 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dcdf1a1-102d-4d4c-adf5-1736b0ea1888" containerName="mariadb-client-1" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.340899 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dcdf1a1-102d-4d4c-adf5-1736b0ea1888" containerName="mariadb-client-1" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.341501 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.343504 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rqp8s" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.354519 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.387958 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnxrk\" (UniqueName: \"kubernetes.io/projected/cfa9fd7f-787e-4587-93e0-ecf50bf46189-kube-api-access-gnxrk\") pod \"mariadb-client-4-default\" (UID: \"cfa9fd7f-787e-4587-93e0-ecf50bf46189\") " pod="openstack/mariadb-client-4-default" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.489461 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnxrk\" (UniqueName: \"kubernetes.io/projected/cfa9fd7f-787e-4587-93e0-ecf50bf46189-kube-api-access-gnxrk\") pod \"mariadb-client-4-default\" (UID: \"cfa9fd7f-787e-4587-93e0-ecf50bf46189\") " pod="openstack/mariadb-client-4-default" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.786924 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnxrk\" (UniqueName: \"kubernetes.io/projected/cfa9fd7f-787e-4587-93e0-ecf50bf46189-kube-api-access-gnxrk\") pod \"mariadb-client-4-default\" (UID: \"cfa9fd7f-787e-4587-93e0-ecf50bf46189\") " pod="openstack/mariadb-client-4-default" Oct 07 23:30:08 crc kubenswrapper[4871]: I1007 23:30:08.958320 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 07 23:30:09 crc kubenswrapper[4871]: I1007 23:30:09.007328 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dcdf1a1-102d-4d4c-adf5-1736b0ea1888" path="/var/lib/kubelet/pods/7dcdf1a1-102d-4d4c-adf5-1736b0ea1888/volumes" Oct 07 23:30:09 crc kubenswrapper[4871]: I1007 23:30:09.504058 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 07 23:30:09 crc kubenswrapper[4871]: W1007 23:30:09.514987 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfa9fd7f_787e_4587_93e0_ecf50bf46189.slice/crio-83a76886481c373f734d93660b88be843085e63c05b27b181260159f7b9c0546 WatchSource:0}: Error finding container 83a76886481c373f734d93660b88be843085e63c05b27b181260159f7b9c0546: Status 404 returned error can't find the container with id 83a76886481c373f734d93660b88be843085e63c05b27b181260159f7b9c0546 Oct 07 23:30:10 crc kubenswrapper[4871]: I1007 23:30:10.306918 4871 generic.go:334] "Generic (PLEG): container finished" podID="cfa9fd7f-787e-4587-93e0-ecf50bf46189" containerID="0ed394a9bfeb0ddda0005ec76d4e535d2cdb22dce1d00920901b1b01fa6ef109" exitCode=0 Oct 07 23:30:10 crc kubenswrapper[4871]: I1007 23:30:10.307031 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"cfa9fd7f-787e-4587-93e0-ecf50bf46189","Type":"ContainerDied","Data":"0ed394a9bfeb0ddda0005ec76d4e535d2cdb22dce1d00920901b1b01fa6ef109"} Oct 07 23:30:10 crc kubenswrapper[4871]: I1007 23:30:10.307363 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"cfa9fd7f-787e-4587-93e0-ecf50bf46189","Type":"ContainerStarted","Data":"83a76886481c373f734d93660b88be843085e63c05b27b181260159f7b9c0546"} Oct 07 23:30:11 crc kubenswrapper[4871]: I1007 23:30:11.740982 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 07 23:30:11 crc kubenswrapper[4871]: I1007 23:30:11.759161 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_cfa9fd7f-787e-4587-93e0-ecf50bf46189/mariadb-client-4-default/0.log" Oct 07 23:30:11 crc kubenswrapper[4871]: I1007 23:30:11.787669 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 07 23:30:11 crc kubenswrapper[4871]: I1007 23:30:11.793223 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 07 23:30:11 crc kubenswrapper[4871]: I1007 23:30:11.839710 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnxrk\" (UniqueName: \"kubernetes.io/projected/cfa9fd7f-787e-4587-93e0-ecf50bf46189-kube-api-access-gnxrk\") pod \"cfa9fd7f-787e-4587-93e0-ecf50bf46189\" (UID: \"cfa9fd7f-787e-4587-93e0-ecf50bf46189\") " Oct 07 23:30:11 crc kubenswrapper[4871]: I1007 23:30:11.847387 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa9fd7f-787e-4587-93e0-ecf50bf46189-kube-api-access-gnxrk" (OuterVolumeSpecName: "kube-api-access-gnxrk") pod "cfa9fd7f-787e-4587-93e0-ecf50bf46189" (UID: "cfa9fd7f-787e-4587-93e0-ecf50bf46189"). InnerVolumeSpecName "kube-api-access-gnxrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:30:11 crc kubenswrapper[4871]: I1007 23:30:11.941270 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnxrk\" (UniqueName: \"kubernetes.io/projected/cfa9fd7f-787e-4587-93e0-ecf50bf46189-kube-api-access-gnxrk\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:12 crc kubenswrapper[4871]: I1007 23:30:12.330992 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83a76886481c373f734d93660b88be843085e63c05b27b181260159f7b9c0546" Oct 07 23:30:12 crc kubenswrapper[4871]: I1007 23:30:12.331090 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 07 23:30:12 crc kubenswrapper[4871]: I1007 23:30:12.994655 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfa9fd7f-787e-4587-93e0-ecf50bf46189" path="/var/lib/kubelet/pods/cfa9fd7f-787e-4587-93e0-ecf50bf46189/volumes" Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.441338 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 07 23:30:16 crc kubenswrapper[4871]: E1007 23:30:16.442112 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa9fd7f-787e-4587-93e0-ecf50bf46189" containerName="mariadb-client-4-default" Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.442159 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa9fd7f-787e-4587-93e0-ecf50bf46189" containerName="mariadb-client-4-default" Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.442589 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfa9fd7f-787e-4587-93e0-ecf50bf46189" containerName="mariadb-client-4-default" Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.443876 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.446906 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rqp8s" Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.454136 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.519148 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7c6t\" (UniqueName: \"kubernetes.io/projected/9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6-kube-api-access-w7c6t\") pod \"mariadb-client-5-default\" (UID: \"9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6\") " pod="openstack/mariadb-client-5-default" Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.621840 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7c6t\" (UniqueName: \"kubernetes.io/projected/9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6-kube-api-access-w7c6t\") pod \"mariadb-client-5-default\" (UID: \"9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6\") " pod="openstack/mariadb-client-5-default" Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.646908 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7c6t\" (UniqueName: \"kubernetes.io/projected/9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6-kube-api-access-w7c6t\") pod \"mariadb-client-5-default\" (UID: \"9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6\") " pod="openstack/mariadb-client-5-default" Oct 07 23:30:16 crc kubenswrapper[4871]: I1007 23:30:16.782535 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 07 23:30:17 crc kubenswrapper[4871]: I1007 23:30:17.213952 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 07 23:30:17 crc kubenswrapper[4871]: I1007 23:30:17.375517 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6","Type":"ContainerStarted","Data":"e5d3409521eadf8bbe363a1702c7711e2b44f3345eb7b6b9bfd746d3d0af3b03"} Oct 07 23:30:18 crc kubenswrapper[4871]: I1007 23:30:18.383614 4871 generic.go:334] "Generic (PLEG): container finished" podID="9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6" containerID="83725a48ff96b0eb167064f5d201a58371e812a6dd6fa3dab6e28e2a8a29b846" exitCode=0 Oct 07 23:30:18 crc kubenswrapper[4871]: I1007 23:30:18.383668 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6","Type":"ContainerDied","Data":"83725a48ff96b0eb167064f5d201a58371e812a6dd6fa3dab6e28e2a8a29b846"} Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.729550 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.748380 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6/mariadb-client-5-default/0.log" Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.776952 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.780865 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7c6t\" (UniqueName: \"kubernetes.io/projected/9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6-kube-api-access-w7c6t\") pod \"9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6\" (UID: \"9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6\") " Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.784954 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.974230 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 07 23:30:19 crc kubenswrapper[4871]: E1007 23:30:19.974833 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6" containerName="mariadb-client-5-default" Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.974848 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6" containerName="mariadb-client-5-default" Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.975004 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6" containerName="mariadb-client-5-default" Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.975469 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 07 23:30:19 crc kubenswrapper[4871]: I1007 23:30:19.984892 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.084857 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg78p\" (UniqueName: \"kubernetes.io/projected/9b412ae3-de69-4587-a284-4a5d845e4aeb-kube-api-access-kg78p\") pod \"mariadb-client-6-default\" (UID: \"9b412ae3-de69-4587-a284-4a5d845e4aeb\") " pod="openstack/mariadb-client-6-default" Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.187302 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg78p\" (UniqueName: \"kubernetes.io/projected/9b412ae3-de69-4587-a284-4a5d845e4aeb-kube-api-access-kg78p\") pod \"mariadb-client-6-default\" (UID: \"9b412ae3-de69-4587-a284-4a5d845e4aeb\") " pod="openstack/mariadb-client-6-default" Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.277110 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6-kube-api-access-w7c6t" (OuterVolumeSpecName: "kube-api-access-w7c6t") pod "9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6" (UID: "9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6"). InnerVolumeSpecName "kube-api-access-w7c6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.278084 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg78p\" (UniqueName: \"kubernetes.io/projected/9b412ae3-de69-4587-a284-4a5d845e4aeb-kube-api-access-kg78p\") pod \"mariadb-client-6-default\" (UID: \"9b412ae3-de69-4587-a284-4a5d845e4aeb\") " pod="openstack/mariadb-client-6-default" Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.288341 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7c6t\" (UniqueName: \"kubernetes.io/projected/9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6-kube-api-access-w7c6t\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.291061 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.405583 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5d3409521eadf8bbe363a1702c7711e2b44f3345eb7b6b9bfd746d3d0af3b03" Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.405662 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.633353 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 07 23:30:20 crc kubenswrapper[4871]: W1007 23:30:20.640908 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b412ae3_de69_4587_a284_4a5d845e4aeb.slice/crio-e44e78ee321ec6b691757dc2ee0783a991e46a5058df92ad9a378ad53b269749 WatchSource:0}: Error finding container e44e78ee321ec6b691757dc2ee0783a991e46a5058df92ad9a378ad53b269749: Status 404 returned error can't find the container with id e44e78ee321ec6b691757dc2ee0783a991e46a5058df92ad9a378ad53b269749 Oct 07 23:30:20 crc kubenswrapper[4871]: I1007 23:30:20.994493 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6" path="/var/lib/kubelet/pods/9a2fb5ff-c84f-411c-a1fd-fc8ad9e8a8d6/volumes" Oct 07 23:30:21 crc kubenswrapper[4871]: I1007 23:30:21.419031 4871 generic.go:334] "Generic (PLEG): container finished" podID="9b412ae3-de69-4587-a284-4a5d845e4aeb" containerID="0be9dc2bc1f5930844773b5fefb20ed7ef9609d1b131bba464d11f5e77a13c0b" exitCode=0 Oct 07 23:30:21 crc kubenswrapper[4871]: I1007 23:30:21.419176 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"9b412ae3-de69-4587-a284-4a5d845e4aeb","Type":"ContainerDied","Data":"0be9dc2bc1f5930844773b5fefb20ed7ef9609d1b131bba464d11f5e77a13c0b"} Oct 07 23:30:21 crc kubenswrapper[4871]: I1007 23:30:21.419665 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"9b412ae3-de69-4587-a284-4a5d845e4aeb","Type":"ContainerStarted","Data":"e44e78ee321ec6b691757dc2ee0783a991e46a5058df92ad9a378ad53b269749"} Oct 07 23:30:21 crc kubenswrapper[4871]: E1007 23:30:21.525244 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b412ae3_de69_4587_a284_4a5d845e4aeb.slice/crio-0be9dc2bc1f5930844773b5fefb20ed7ef9609d1b131bba464d11f5e77a13c0b.scope\": RecentStats: unable to find data in memory cache]" Oct 07 23:30:22 crc kubenswrapper[4871]: I1007 23:30:22.929372 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 07 23:30:22 crc kubenswrapper[4871]: I1007 23:30:22.962774 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_9b412ae3-de69-4587-a284-4a5d845e4aeb/mariadb-client-6-default/0.log" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.006811 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.011544 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.132343 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg78p\" (UniqueName: \"kubernetes.io/projected/9b412ae3-de69-4587-a284-4a5d845e4aeb-kube-api-access-kg78p\") pod \"9b412ae3-de69-4587-a284-4a5d845e4aeb\" (UID: \"9b412ae3-de69-4587-a284-4a5d845e4aeb\") " Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.141231 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b412ae3-de69-4587-a284-4a5d845e4aeb-kube-api-access-kg78p" (OuterVolumeSpecName: "kube-api-access-kg78p") pod "9b412ae3-de69-4587-a284-4a5d845e4aeb" (UID: "9b412ae3-de69-4587-a284-4a5d845e4aeb"). InnerVolumeSpecName "kube-api-access-kg78p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.190132 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 07 23:30:23 crc kubenswrapper[4871]: E1007 23:30:23.190485 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b412ae3-de69-4587-a284-4a5d845e4aeb" containerName="mariadb-client-6-default" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.190501 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b412ae3-de69-4587-a284-4a5d845e4aeb" containerName="mariadb-client-6-default" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.190693 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b412ae3-de69-4587-a284-4a5d845e4aeb" containerName="mariadb-client-6-default" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.191345 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.208200 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.234515 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg78p\" (UniqueName: \"kubernetes.io/projected/9b412ae3-de69-4587-a284-4a5d845e4aeb-kube-api-access-kg78p\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.335856 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-294bd\" (UniqueName: \"kubernetes.io/projected/f0e5ed80-1d29-4978-9932-9e02e239192d-kube-api-access-294bd\") pod \"mariadb-client-7-default\" (UID: \"f0e5ed80-1d29-4978-9932-9e02e239192d\") " pod="openstack/mariadb-client-7-default" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.436950 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-294bd\" (UniqueName: \"kubernetes.io/projected/f0e5ed80-1d29-4978-9932-9e02e239192d-kube-api-access-294bd\") pod \"mariadb-client-7-default\" (UID: \"f0e5ed80-1d29-4978-9932-9e02e239192d\") " pod="openstack/mariadb-client-7-default" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.443565 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e44e78ee321ec6b691757dc2ee0783a991e46a5058df92ad9a378ad53b269749" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.443666 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.458925 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-294bd\" (UniqueName: \"kubernetes.io/projected/f0e5ed80-1d29-4978-9932-9e02e239192d-kube-api-access-294bd\") pod \"mariadb-client-7-default\" (UID: \"f0e5ed80-1d29-4978-9932-9e02e239192d\") " pod="openstack/mariadb-client-7-default" Oct 07 23:30:23 crc kubenswrapper[4871]: I1007 23:30:23.516189 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 07 23:30:24 crc kubenswrapper[4871]: W1007 23:30:24.042011 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0e5ed80_1d29_4978_9932_9e02e239192d.slice/crio-b813ae4070c773c8ef15fd51cc798e7d97beee3f2cec2fda57e3129dc785d95e WatchSource:0}: Error finding container b813ae4070c773c8ef15fd51cc798e7d97beee3f2cec2fda57e3129dc785d95e: Status 404 returned error can't find the container with id b813ae4070c773c8ef15fd51cc798e7d97beee3f2cec2fda57e3129dc785d95e Oct 07 23:30:24 crc kubenswrapper[4871]: I1007 23:30:24.052327 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 07 23:30:24 crc kubenswrapper[4871]: I1007 23:30:24.458641 4871 generic.go:334] "Generic (PLEG): container finished" podID="f0e5ed80-1d29-4978-9932-9e02e239192d" containerID="395c971e87ea1dcd6f5b65a51c13b1b9fad8615c039fe09d30696e4461bf8625" exitCode=0 Oct 07 23:30:24 crc kubenswrapper[4871]: I1007 23:30:24.458730 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"f0e5ed80-1d29-4978-9932-9e02e239192d","Type":"ContainerDied","Data":"395c971e87ea1dcd6f5b65a51c13b1b9fad8615c039fe09d30696e4461bf8625"} Oct 07 23:30:24 crc kubenswrapper[4871]: I1007 23:30:24.458780 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"f0e5ed80-1d29-4978-9932-9e02e239192d","Type":"ContainerStarted","Data":"b813ae4070c773c8ef15fd51cc798e7d97beee3f2cec2fda57e3129dc785d95e"} Oct 07 23:30:24 crc kubenswrapper[4871]: I1007 23:30:24.997999 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b412ae3-de69-4587-a284-4a5d845e4aeb" path="/var/lib/kubelet/pods/9b412ae3-de69-4587-a284-4a5d845e4aeb/volumes" Oct 07 23:30:25 crc kubenswrapper[4871]: I1007 23:30:25.844997 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 07 23:30:25 crc kubenswrapper[4871]: I1007 23:30:25.860020 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_f0e5ed80-1d29-4978-9932-9e02e239192d/mariadb-client-7-default/0.log" Oct 07 23:30:25 crc kubenswrapper[4871]: I1007 23:30:25.878712 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-294bd\" (UniqueName: \"kubernetes.io/projected/f0e5ed80-1d29-4978-9932-9e02e239192d-kube-api-access-294bd\") pod \"f0e5ed80-1d29-4978-9932-9e02e239192d\" (UID: \"f0e5ed80-1d29-4978-9932-9e02e239192d\") " Oct 07 23:30:25 crc kubenswrapper[4871]: I1007 23:30:25.879977 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 07 23:30:25 crc kubenswrapper[4871]: I1007 23:30:25.884356 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 07 23:30:25 crc kubenswrapper[4871]: I1007 23:30:25.885759 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0e5ed80-1d29-4978-9932-9e02e239192d-kube-api-access-294bd" (OuterVolumeSpecName: "kube-api-access-294bd") pod "f0e5ed80-1d29-4978-9932-9e02e239192d" (UID: "f0e5ed80-1d29-4978-9932-9e02e239192d"). InnerVolumeSpecName "kube-api-access-294bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:30:25 crc kubenswrapper[4871]: I1007 23:30:25.981245 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-294bd\" (UniqueName: \"kubernetes.io/projected/f0e5ed80-1d29-4978-9932-9e02e239192d-kube-api-access-294bd\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.007787 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 07 23:30:26 crc kubenswrapper[4871]: E1007 23:30:26.009084 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0e5ed80-1d29-4978-9932-9e02e239192d" containerName="mariadb-client-7-default" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.009254 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0e5ed80-1d29-4978-9932-9e02e239192d" containerName="mariadb-client-7-default" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.009669 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0e5ed80-1d29-4978-9932-9e02e239192d" containerName="mariadb-client-7-default" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.011642 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.014232 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.185651 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnh7p\" (UniqueName: \"kubernetes.io/projected/09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1-kube-api-access-nnh7p\") pod \"mariadb-client-2\" (UID: \"09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1\") " pod="openstack/mariadb-client-2" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.287688 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnh7p\" (UniqueName: \"kubernetes.io/projected/09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1-kube-api-access-nnh7p\") pod \"mariadb-client-2\" (UID: \"09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1\") " pod="openstack/mariadb-client-2" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.321287 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnh7p\" (UniqueName: \"kubernetes.io/projected/09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1-kube-api-access-nnh7p\") pod \"mariadb-client-2\" (UID: \"09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1\") " pod="openstack/mariadb-client-2" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.351333 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.487720 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b813ae4070c773c8ef15fd51cc798e7d97beee3f2cec2fda57e3129dc785d95e" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.487826 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.892011 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 07 23:30:26 crc kubenswrapper[4871]: W1007 23:30:26.894569 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09721ccb_bdfe_4d0e_8ab9_5cff7c4a86c1.slice/crio-a38bae8db92cb27d01db02c21d7322123e2d1bb9c7b4b651945588d433db1172 WatchSource:0}: Error finding container a38bae8db92cb27d01db02c21d7322123e2d1bb9c7b4b651945588d433db1172: Status 404 returned error can't find the container with id a38bae8db92cb27d01db02c21d7322123e2d1bb9c7b4b651945588d433db1172 Oct 07 23:30:26 crc kubenswrapper[4871]: I1007 23:30:26.993535 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0e5ed80-1d29-4978-9932-9e02e239192d" path="/var/lib/kubelet/pods/f0e5ed80-1d29-4978-9932-9e02e239192d/volumes" Oct 07 23:30:27 crc kubenswrapper[4871]: I1007 23:30:27.500032 4871 generic.go:334] "Generic (PLEG): container finished" podID="09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1" containerID="dbac19c9f543e76ee7d9dece2daaf599ffdd5963c4dd867ef1e9421c428b7000" exitCode=0 Oct 07 23:30:27 crc kubenswrapper[4871]: I1007 23:30:27.500230 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1","Type":"ContainerDied","Data":"dbac19c9f543e76ee7d9dece2daaf599ffdd5963c4dd867ef1e9421c428b7000"} Oct 07 23:30:27 crc kubenswrapper[4871]: I1007 23:30:27.500466 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1","Type":"ContainerStarted","Data":"a38bae8db92cb27d01db02c21d7322123e2d1bb9c7b4b651945588d433db1172"} Oct 07 23:30:28 crc kubenswrapper[4871]: I1007 23:30:28.976146 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 07 23:30:29 crc kubenswrapper[4871]: I1007 23:30:29.008151 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1/mariadb-client-2/0.log" Oct 07 23:30:29 crc kubenswrapper[4871]: I1007 23:30:29.040424 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 07 23:30:29 crc kubenswrapper[4871]: I1007 23:30:29.047980 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 07 23:30:29 crc kubenswrapper[4871]: I1007 23:30:29.057550 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnh7p\" (UniqueName: \"kubernetes.io/projected/09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1-kube-api-access-nnh7p\") pod \"09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1\" (UID: \"09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1\") " Oct 07 23:30:29 crc kubenswrapper[4871]: I1007 23:30:29.071177 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1-kube-api-access-nnh7p" (OuterVolumeSpecName: "kube-api-access-nnh7p") pod "09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1" (UID: "09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1"). InnerVolumeSpecName "kube-api-access-nnh7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:30:29 crc kubenswrapper[4871]: I1007 23:30:29.159385 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnh7p\" (UniqueName: \"kubernetes.io/projected/09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1-kube-api-access-nnh7p\") on node \"crc\" DevicePath \"\"" Oct 07 23:30:29 crc kubenswrapper[4871]: I1007 23:30:29.522438 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a38bae8db92cb27d01db02c21d7322123e2d1bb9c7b4b651945588d433db1172" Oct 07 23:30:29 crc kubenswrapper[4871]: I1007 23:30:29.522532 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 07 23:30:30 crc kubenswrapper[4871]: I1007 23:30:30.996648 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1" path="/var/lib/kubelet/pods/09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1/volumes" Oct 07 23:30:55 crc kubenswrapper[4871]: I1007 23:30:55.979065 4871 scope.go:117] "RemoveContainer" containerID="b386ef48a4671a721e35166bc7d37f21fb26ddf6426a8817770002609981cab5" Oct 07 23:30:56 crc kubenswrapper[4871]: I1007 23:30:56.009935 4871 scope.go:117] "RemoveContainer" containerID="0c96262bf0a814a6f26dce4a1cf750095780f028692d0eedfcfefb7ffcaf3d79" Oct 07 23:31:05 crc kubenswrapper[4871]: I1007 23:31:05.514631 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:31:05 crc kubenswrapper[4871]: I1007 23:31:05.516377 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:31:35 crc kubenswrapper[4871]: I1007 23:31:35.512160 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:31:35 crc kubenswrapper[4871]: I1007 23:31:35.512755 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:32:05 crc kubenswrapper[4871]: I1007 23:32:05.512365 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:32:05 crc kubenswrapper[4871]: I1007 23:32:05.513407 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:32:05 crc kubenswrapper[4871]: I1007 23:32:05.513486 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:32:05 crc kubenswrapper[4871]: I1007 23:32:05.514542 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:32:05 crc kubenswrapper[4871]: I1007 23:32:05.514647 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" gracePeriod=600 Oct 07 23:32:05 crc kubenswrapper[4871]: E1007 23:32:05.649783 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:32:06 crc kubenswrapper[4871]: I1007 23:32:06.501736 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" exitCode=0 Oct 07 23:32:06 crc kubenswrapper[4871]: I1007 23:32:06.501846 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a"} Oct 07 23:32:06 crc kubenswrapper[4871]: I1007 23:32:06.501979 4871 scope.go:117] "RemoveContainer" containerID="3ab5d945d29b218b1fcfe5ad517fbc83535bc60079d91b4857bc595f89a8605f" Oct 07 23:32:06 crc kubenswrapper[4871]: I1007 23:32:06.502757 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:32:06 crc kubenswrapper[4871]: E1007 23:32:06.503097 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:32:17 crc kubenswrapper[4871]: I1007 23:32:17.982299 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:32:17 crc kubenswrapper[4871]: E1007 23:32:17.983565 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:32:32 crc kubenswrapper[4871]: I1007 23:32:32.982186 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:32:32 crc kubenswrapper[4871]: E1007 23:32:32.983300 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:32:45 crc kubenswrapper[4871]: I1007 23:32:45.982872 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:32:45 crc kubenswrapper[4871]: E1007 23:32:45.983837 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:32:58 crc kubenswrapper[4871]: I1007 23:32:58.982256 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:32:58 crc kubenswrapper[4871]: E1007 23:32:58.983224 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:33:10 crc kubenswrapper[4871]: I1007 23:33:10.982862 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:33:10 crc kubenswrapper[4871]: E1007 23:33:10.984224 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.247926 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t2525"] Oct 07 23:33:16 crc kubenswrapper[4871]: E1007 23:33:16.248708 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1" containerName="mariadb-client-2" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.248722 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1" containerName="mariadb-client-2" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.248882 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="09721ccb-bdfe-4d0e-8ab9-5cff7c4a86c1" containerName="mariadb-client-2" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.249919 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.264134 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2525"] Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.326010 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-utilities\") pod \"certified-operators-t2525\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.326053 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xvzf\" (UniqueName: \"kubernetes.io/projected/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-kube-api-access-5xvzf\") pod \"certified-operators-t2525\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.326373 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-catalog-content\") pod \"certified-operators-t2525\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.427598 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-catalog-content\") pod \"certified-operators-t2525\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.427924 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-utilities\") pod \"certified-operators-t2525\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.427989 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xvzf\" (UniqueName: \"kubernetes.io/projected/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-kube-api-access-5xvzf\") pod \"certified-operators-t2525\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.428436 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-catalog-content\") pod \"certified-operators-t2525\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.428648 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-utilities\") pod \"certified-operators-t2525\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.453727 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xvzf\" (UniqueName: \"kubernetes.io/projected/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-kube-api-access-5xvzf\") pod \"certified-operators-t2525\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:16 crc kubenswrapper[4871]: I1007 23:33:16.582095 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:17 crc kubenswrapper[4871]: I1007 23:33:17.042050 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2525"] Oct 07 23:33:18 crc kubenswrapper[4871]: I1007 23:33:18.216835 4871 generic.go:334] "Generic (PLEG): container finished" podID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerID="62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718" exitCode=0 Oct 07 23:33:18 crc kubenswrapper[4871]: I1007 23:33:18.217153 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2525" event={"ID":"317f3cfe-1b1e-4bac-92dc-10e77c51ee72","Type":"ContainerDied","Data":"62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718"} Oct 07 23:33:18 crc kubenswrapper[4871]: I1007 23:33:18.217185 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2525" event={"ID":"317f3cfe-1b1e-4bac-92dc-10e77c51ee72","Type":"ContainerStarted","Data":"bee10c79d33beea932d6547a67a02eca069065668a87453b6d808a98b3168999"} Oct 07 23:33:20 crc kubenswrapper[4871]: I1007 23:33:20.238542 4871 generic.go:334] "Generic (PLEG): container finished" podID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerID="7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad" exitCode=0 Oct 07 23:33:20 crc kubenswrapper[4871]: I1007 23:33:20.238722 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2525" event={"ID":"317f3cfe-1b1e-4bac-92dc-10e77c51ee72","Type":"ContainerDied","Data":"7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad"} Oct 07 23:33:21 crc kubenswrapper[4871]: I1007 23:33:21.250893 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2525" event={"ID":"317f3cfe-1b1e-4bac-92dc-10e77c51ee72","Type":"ContainerStarted","Data":"81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee"} Oct 07 23:33:21 crc kubenswrapper[4871]: I1007 23:33:21.276544 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t2525" podStartSLOduration=2.812204098 podStartE2EDuration="5.276515953s" podCreationTimestamp="2025-10-07 23:33:16 +0000 UTC" firstStartedPulling="2025-10-07 23:33:18.218240067 +0000 UTC m=+5072.020938140" lastFinishedPulling="2025-10-07 23:33:20.682551882 +0000 UTC m=+5074.485249995" observedRunningTime="2025-10-07 23:33:21.270896944 +0000 UTC m=+5075.073595077" watchObservedRunningTime="2025-10-07 23:33:21.276515953 +0000 UTC m=+5075.079214066" Oct 07 23:33:24 crc kubenswrapper[4871]: I1007 23:33:24.983107 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:33:24 crc kubenswrapper[4871]: E1007 23:33:24.983787 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:33:26 crc kubenswrapper[4871]: I1007 23:33:26.582867 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:26 crc kubenswrapper[4871]: I1007 23:33:26.583294 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:26 crc kubenswrapper[4871]: I1007 23:33:26.664950 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:27 crc kubenswrapper[4871]: I1007 23:33:27.406041 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:27 crc kubenswrapper[4871]: I1007 23:33:27.483492 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2525"] Oct 07 23:33:29 crc kubenswrapper[4871]: I1007 23:33:29.337977 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t2525" podUID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerName="registry-server" containerID="cri-o://81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee" gracePeriod=2 Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.299265 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.346439 4871 generic.go:334] "Generic (PLEG): container finished" podID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerID="81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee" exitCode=0 Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.346498 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2525" event={"ID":"317f3cfe-1b1e-4bac-92dc-10e77c51ee72","Type":"ContainerDied","Data":"81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee"} Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.346533 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2525" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.346581 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2525" event={"ID":"317f3cfe-1b1e-4bac-92dc-10e77c51ee72","Type":"ContainerDied","Data":"bee10c79d33beea932d6547a67a02eca069065668a87453b6d808a98b3168999"} Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.346626 4871 scope.go:117] "RemoveContainer" containerID="81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.373134 4871 scope.go:117] "RemoveContainer" containerID="7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.394316 4871 scope.go:117] "RemoveContainer" containerID="62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.395300 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-utilities\") pod \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.395371 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-catalog-content\") pod \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.395579 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xvzf\" (UniqueName: \"kubernetes.io/projected/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-kube-api-access-5xvzf\") pod \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\" (UID: \"317f3cfe-1b1e-4bac-92dc-10e77c51ee72\") " Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.397218 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-utilities" (OuterVolumeSpecName: "utilities") pod "317f3cfe-1b1e-4bac-92dc-10e77c51ee72" (UID: "317f3cfe-1b1e-4bac-92dc-10e77c51ee72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.405298 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-kube-api-access-5xvzf" (OuterVolumeSpecName: "kube-api-access-5xvzf") pod "317f3cfe-1b1e-4bac-92dc-10e77c51ee72" (UID: "317f3cfe-1b1e-4bac-92dc-10e77c51ee72"). InnerVolumeSpecName "kube-api-access-5xvzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.458154 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "317f3cfe-1b1e-4bac-92dc-10e77c51ee72" (UID: "317f3cfe-1b1e-4bac-92dc-10e77c51ee72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.461602 4871 scope.go:117] "RemoveContainer" containerID="81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee" Oct 07 23:33:30 crc kubenswrapper[4871]: E1007 23:33:30.462197 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee\": container with ID starting with 81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee not found: ID does not exist" containerID="81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.462228 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee"} err="failed to get container status \"81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee\": rpc error: code = NotFound desc = could not find container \"81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee\": container with ID starting with 81d27bdcc21efca0e7ccbb837e712bfce734d796056eb0ee870538b6baccdfee not found: ID does not exist" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.462246 4871 scope.go:117] "RemoveContainer" containerID="7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad" Oct 07 23:33:30 crc kubenswrapper[4871]: E1007 23:33:30.462506 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad\": container with ID starting with 7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad not found: ID does not exist" containerID="7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.462526 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad"} err="failed to get container status \"7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad\": rpc error: code = NotFound desc = could not find container \"7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad\": container with ID starting with 7f6d747c0028e6866768e72a1c1d4189e287a7f98a4a8ae3d4cbaa57f6f5b2ad not found: ID does not exist" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.462537 4871 scope.go:117] "RemoveContainer" containerID="62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718" Oct 07 23:33:30 crc kubenswrapper[4871]: E1007 23:33:30.462811 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718\": container with ID starting with 62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718 not found: ID does not exist" containerID="62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.462833 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718"} err="failed to get container status \"62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718\": rpc error: code = NotFound desc = could not find container \"62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718\": container with ID starting with 62dd30b2d4c9faac24b5d480d9b1a81aec2c3e5838994ef40a8d4830b7c4d718 not found: ID does not exist" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.497284 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.497313 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.497324 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xvzf\" (UniqueName: \"kubernetes.io/projected/317f3cfe-1b1e-4bac-92dc-10e77c51ee72-kube-api-access-5xvzf\") on node \"crc\" DevicePath \"\"" Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.701170 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2525"] Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.707188 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t2525"] Oct 07 23:33:30 crc kubenswrapper[4871]: I1007 23:33:30.995914 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" path="/var/lib/kubelet/pods/317f3cfe-1b1e-4bac-92dc-10e77c51ee72/volumes" Oct 07 23:33:37 crc kubenswrapper[4871]: I1007 23:33:37.983697 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:33:37 crc kubenswrapper[4871]: E1007 23:33:37.984787 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:33:51 crc kubenswrapper[4871]: I1007 23:33:51.982391 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:33:51 crc kubenswrapper[4871]: E1007 23:33:51.983378 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:33:56 crc kubenswrapper[4871]: I1007 23:33:56.172667 4871 scope.go:117] "RemoveContainer" containerID="a5288a75ee8a01032fe8be52619292a8b1261515d2c65c655e9b313a4c3e34c3" Oct 07 23:34:06 crc kubenswrapper[4871]: I1007 23:34:06.990900 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:34:06 crc kubenswrapper[4871]: E1007 23:34:06.991899 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:34:17 crc kubenswrapper[4871]: I1007 23:34:17.983251 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:34:17 crc kubenswrapper[4871]: E1007 23:34:17.984559 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:34:29 crc kubenswrapper[4871]: I1007 23:34:29.982253 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:34:29 crc kubenswrapper[4871]: E1007 23:34:29.983233 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:34:42 crc kubenswrapper[4871]: I1007 23:34:42.983175 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:34:42 crc kubenswrapper[4871]: E1007 23:34:42.984163 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:34:55 crc kubenswrapper[4871]: I1007 23:34:55.982980 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:34:55 crc kubenswrapper[4871]: E1007 23:34:55.984285 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:34:56 crc kubenswrapper[4871]: I1007 23:34:56.237320 4871 scope.go:117] "RemoveContainer" containerID="0c93463ceb81ee6ad5efa72352bcfcf172b81eec903295446b2a3e57b4569122" Oct 07 23:34:56 crc kubenswrapper[4871]: I1007 23:34:56.264558 4871 scope.go:117] "RemoveContainer" containerID="787ce125fd21c4f80f1cec52d7862b9d28e91647a285000e0cb5a7aa35118a12" Oct 07 23:34:56 crc kubenswrapper[4871]: I1007 23:34:56.297877 4871 scope.go:117] "RemoveContainer" containerID="9405df51dcb2c3792b3e3ab7a0987a7a9c5f649ce8734b90f2274fb13a390cd3" Oct 07 23:34:56 crc kubenswrapper[4871]: I1007 23:34:56.346941 4871 scope.go:117] "RemoveContainer" containerID="3d3dad8bd943cffdec74358f92205f5ef6e3354d6f169696b22c4f86162de8bb" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.280266 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 07 23:35:01 crc kubenswrapper[4871]: E1007 23:35:01.281147 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerName="registry-server" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.281165 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerName="registry-server" Oct 07 23:35:01 crc kubenswrapper[4871]: E1007 23:35:01.281184 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerName="extract-utilities" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.281192 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerName="extract-utilities" Oct 07 23:35:01 crc kubenswrapper[4871]: E1007 23:35:01.281211 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerName="extract-content" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.281219 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerName="extract-content" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.281400 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="317f3cfe-1b1e-4bac-92dc-10e77c51ee72" containerName="registry-server" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.282035 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.286596 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rqp8s" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.290293 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.376330 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\") pod \"mariadb-copy-data\" (UID: \"2c49b518-7037-4062-99ad-072d766d43a3\") " pod="openstack/mariadb-copy-data" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.376414 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r22j\" (UniqueName: \"kubernetes.io/projected/2c49b518-7037-4062-99ad-072d766d43a3-kube-api-access-2r22j\") pod \"mariadb-copy-data\" (UID: \"2c49b518-7037-4062-99ad-072d766d43a3\") " pod="openstack/mariadb-copy-data" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.478047 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\") pod \"mariadb-copy-data\" (UID: \"2c49b518-7037-4062-99ad-072d766d43a3\") " pod="openstack/mariadb-copy-data" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.478117 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r22j\" (UniqueName: \"kubernetes.io/projected/2c49b518-7037-4062-99ad-072d766d43a3-kube-api-access-2r22j\") pod \"mariadb-copy-data\" (UID: \"2c49b518-7037-4062-99ad-072d766d43a3\") " pod="openstack/mariadb-copy-data" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.480294 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.480569 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\") pod \"mariadb-copy-data\" (UID: \"2c49b518-7037-4062-99ad-072d766d43a3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a81d935b230dd09634403ca41a81a528136a1e345700226661c8e99f89a9ded4/globalmount\"" pod="openstack/mariadb-copy-data" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.497686 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r22j\" (UniqueName: \"kubernetes.io/projected/2c49b518-7037-4062-99ad-072d766d43a3-kube-api-access-2r22j\") pod \"mariadb-copy-data\" (UID: \"2c49b518-7037-4062-99ad-072d766d43a3\") " pod="openstack/mariadb-copy-data" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.514936 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\") pod \"mariadb-copy-data\" (UID: \"2c49b518-7037-4062-99ad-072d766d43a3\") " pod="openstack/mariadb-copy-data" Oct 07 23:35:01 crc kubenswrapper[4871]: I1007 23:35:01.604112 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 07 23:35:02 crc kubenswrapper[4871]: I1007 23:35:02.116094 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 07 23:35:02 crc kubenswrapper[4871]: I1007 23:35:02.334707 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"2c49b518-7037-4062-99ad-072d766d43a3","Type":"ContainerStarted","Data":"b1c6e628fd76c5e078863763c51ee084754bb5149b6e6a64e6ae2987a7d1fc8f"} Oct 07 23:35:03 crc kubenswrapper[4871]: I1007 23:35:03.343191 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"2c49b518-7037-4062-99ad-072d766d43a3","Type":"ContainerStarted","Data":"ba2c3e5ec3b1e578580bf7573002682138386e24086f6b08f080ce48de024efa"} Oct 07 23:35:03 crc kubenswrapper[4871]: I1007 23:35:03.364498 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.3644788549999998 podStartE2EDuration="3.364478855s" podCreationTimestamp="2025-10-07 23:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:35:03.359819472 +0000 UTC m=+5177.162517565" watchObservedRunningTime="2025-10-07 23:35:03.364478855 +0000 UTC m=+5177.167176938" Oct 07 23:35:05 crc kubenswrapper[4871]: I1007 23:35:05.292064 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:05 crc kubenswrapper[4871]: I1007 23:35:05.293044 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 07 23:35:05 crc kubenswrapper[4871]: I1007 23:35:05.304992 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:05 crc kubenswrapper[4871]: I1007 23:35:05.441705 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kbfd\" (UniqueName: \"kubernetes.io/projected/32728537-6340-4110-8875-71b353b5fe08-kube-api-access-4kbfd\") pod \"mariadb-client\" (UID: \"32728537-6340-4110-8875-71b353b5fe08\") " pod="openstack/mariadb-client" Oct 07 23:35:05 crc kubenswrapper[4871]: I1007 23:35:05.542993 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kbfd\" (UniqueName: \"kubernetes.io/projected/32728537-6340-4110-8875-71b353b5fe08-kube-api-access-4kbfd\") pod \"mariadb-client\" (UID: \"32728537-6340-4110-8875-71b353b5fe08\") " pod="openstack/mariadb-client" Oct 07 23:35:05 crc kubenswrapper[4871]: I1007 23:35:05.568911 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kbfd\" (UniqueName: \"kubernetes.io/projected/32728537-6340-4110-8875-71b353b5fe08-kube-api-access-4kbfd\") pod \"mariadb-client\" (UID: \"32728537-6340-4110-8875-71b353b5fe08\") " pod="openstack/mariadb-client" Oct 07 23:35:05 crc kubenswrapper[4871]: I1007 23:35:05.623478 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 07 23:35:06 crc kubenswrapper[4871]: I1007 23:35:06.120244 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:06 crc kubenswrapper[4871]: I1007 23:35:06.379396 4871 generic.go:334] "Generic (PLEG): container finished" podID="32728537-6340-4110-8875-71b353b5fe08" containerID="f1c3dc602ed9c44b90b1e1a8c12ad3a64d6d3fed18b9726a5c3e5cf945ef6447" exitCode=0 Oct 07 23:35:06 crc kubenswrapper[4871]: I1007 23:35:06.379439 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"32728537-6340-4110-8875-71b353b5fe08","Type":"ContainerDied","Data":"f1c3dc602ed9c44b90b1e1a8c12ad3a64d6d3fed18b9726a5c3e5cf945ef6447"} Oct 07 23:35:06 crc kubenswrapper[4871]: I1007 23:35:06.379462 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"32728537-6340-4110-8875-71b353b5fe08","Type":"ContainerStarted","Data":"abff7a54fb15dd57f3352fb015052f33a9f84e08a92756fcff287c4a6e29e329"} Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.752043 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.775599 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_32728537-6340-4110-8875-71b353b5fe08/mariadb-client/0.log" Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.800671 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.805632 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.897263 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kbfd\" (UniqueName: \"kubernetes.io/projected/32728537-6340-4110-8875-71b353b5fe08-kube-api-access-4kbfd\") pod \"32728537-6340-4110-8875-71b353b5fe08\" (UID: \"32728537-6340-4110-8875-71b353b5fe08\") " Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.903368 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32728537-6340-4110-8875-71b353b5fe08-kube-api-access-4kbfd" (OuterVolumeSpecName: "kube-api-access-4kbfd") pod "32728537-6340-4110-8875-71b353b5fe08" (UID: "32728537-6340-4110-8875-71b353b5fe08"). InnerVolumeSpecName "kube-api-access-4kbfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.934930 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:07 crc kubenswrapper[4871]: E1007 23:35:07.935399 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32728537-6340-4110-8875-71b353b5fe08" containerName="mariadb-client" Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.935430 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="32728537-6340-4110-8875-71b353b5fe08" containerName="mariadb-client" Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.935738 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="32728537-6340-4110-8875-71b353b5fe08" containerName="mariadb-client" Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.936530 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.941274 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.999502 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2b8q\" (UniqueName: \"kubernetes.io/projected/3c37ba1e-a7b1-4398-8193-5396701840ea-kube-api-access-w2b8q\") pod \"mariadb-client\" (UID: \"3c37ba1e-a7b1-4398-8193-5396701840ea\") " pod="openstack/mariadb-client" Oct 07 23:35:07 crc kubenswrapper[4871]: I1007 23:35:07.999787 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kbfd\" (UniqueName: \"kubernetes.io/projected/32728537-6340-4110-8875-71b353b5fe08-kube-api-access-4kbfd\") on node \"crc\" DevicePath \"\"" Oct 07 23:35:08 crc kubenswrapper[4871]: I1007 23:35:08.101588 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2b8q\" (UniqueName: \"kubernetes.io/projected/3c37ba1e-a7b1-4398-8193-5396701840ea-kube-api-access-w2b8q\") pod \"mariadb-client\" (UID: \"3c37ba1e-a7b1-4398-8193-5396701840ea\") " pod="openstack/mariadb-client" Oct 07 23:35:08 crc kubenswrapper[4871]: I1007 23:35:08.121570 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2b8q\" (UniqueName: \"kubernetes.io/projected/3c37ba1e-a7b1-4398-8193-5396701840ea-kube-api-access-w2b8q\") pod \"mariadb-client\" (UID: \"3c37ba1e-a7b1-4398-8193-5396701840ea\") " pod="openstack/mariadb-client" Oct 07 23:35:08 crc kubenswrapper[4871]: I1007 23:35:08.276452 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 07 23:35:08 crc kubenswrapper[4871]: I1007 23:35:08.400478 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abff7a54fb15dd57f3352fb015052f33a9f84e08a92756fcff287c4a6e29e329" Oct 07 23:35:08 crc kubenswrapper[4871]: I1007 23:35:08.400532 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 07 23:35:08 crc kubenswrapper[4871]: I1007 23:35:08.434034 4871 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="32728537-6340-4110-8875-71b353b5fe08" podUID="3c37ba1e-a7b1-4398-8193-5396701840ea" Oct 07 23:35:08 crc kubenswrapper[4871]: I1007 23:35:08.762918 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:08 crc kubenswrapper[4871]: W1007 23:35:08.767696 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c37ba1e_a7b1_4398_8193_5396701840ea.slice/crio-8cbb27636080a42544f65eae172f9126a773af6764499f2b4346254ad4f77c90 WatchSource:0}: Error finding container 8cbb27636080a42544f65eae172f9126a773af6764499f2b4346254ad4f77c90: Status 404 returned error can't find the container with id 8cbb27636080a42544f65eae172f9126a773af6764499f2b4346254ad4f77c90 Oct 07 23:35:08 crc kubenswrapper[4871]: I1007 23:35:08.989988 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32728537-6340-4110-8875-71b353b5fe08" path="/var/lib/kubelet/pods/32728537-6340-4110-8875-71b353b5fe08/volumes" Oct 07 23:35:09 crc kubenswrapper[4871]: I1007 23:35:09.412751 4871 generic.go:334] "Generic (PLEG): container finished" podID="3c37ba1e-a7b1-4398-8193-5396701840ea" containerID="1a7dc44799b83de094dcb7f36052fc8131ea02ecb2c592d2907078f31fb6fd3e" exitCode=0 Oct 07 23:35:09 crc kubenswrapper[4871]: I1007 23:35:09.412827 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"3c37ba1e-a7b1-4398-8193-5396701840ea","Type":"ContainerDied","Data":"1a7dc44799b83de094dcb7f36052fc8131ea02ecb2c592d2907078f31fb6fd3e"} Oct 07 23:35:09 crc kubenswrapper[4871]: I1007 23:35:09.412854 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"3c37ba1e-a7b1-4398-8193-5396701840ea","Type":"ContainerStarted","Data":"8cbb27636080a42544f65eae172f9126a773af6764499f2b4346254ad4f77c90"} Oct 07 23:35:09 crc kubenswrapper[4871]: I1007 23:35:09.982412 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:35:09 crc kubenswrapper[4871]: E1007 23:35:09.982851 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:35:10 crc kubenswrapper[4871]: I1007 23:35:10.800465 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 07 23:35:10 crc kubenswrapper[4871]: I1007 23:35:10.822650 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_3c37ba1e-a7b1-4398-8193-5396701840ea/mariadb-client/0.log" Oct 07 23:35:10 crc kubenswrapper[4871]: I1007 23:35:10.847088 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:10 crc kubenswrapper[4871]: I1007 23:35:10.852773 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 07 23:35:10 crc kubenswrapper[4871]: I1007 23:35:10.971243 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2b8q\" (UniqueName: \"kubernetes.io/projected/3c37ba1e-a7b1-4398-8193-5396701840ea-kube-api-access-w2b8q\") pod \"3c37ba1e-a7b1-4398-8193-5396701840ea\" (UID: \"3c37ba1e-a7b1-4398-8193-5396701840ea\") " Oct 07 23:35:10 crc kubenswrapper[4871]: I1007 23:35:10.979458 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c37ba1e-a7b1-4398-8193-5396701840ea-kube-api-access-w2b8q" (OuterVolumeSpecName: "kube-api-access-w2b8q") pod "3c37ba1e-a7b1-4398-8193-5396701840ea" (UID: "3c37ba1e-a7b1-4398-8193-5396701840ea"). InnerVolumeSpecName "kube-api-access-w2b8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:35:10 crc kubenswrapper[4871]: I1007 23:35:10.992984 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c37ba1e-a7b1-4398-8193-5396701840ea" path="/var/lib/kubelet/pods/3c37ba1e-a7b1-4398-8193-5396701840ea/volumes" Oct 07 23:35:11 crc kubenswrapper[4871]: I1007 23:35:11.073041 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2b8q\" (UniqueName: \"kubernetes.io/projected/3c37ba1e-a7b1-4398-8193-5396701840ea-kube-api-access-w2b8q\") on node \"crc\" DevicePath \"\"" Oct 07 23:35:11 crc kubenswrapper[4871]: I1007 23:35:11.432715 4871 scope.go:117] "RemoveContainer" containerID="1a7dc44799b83de094dcb7f36052fc8131ea02ecb2c592d2907078f31fb6fd3e" Oct 07 23:35:11 crc kubenswrapper[4871]: I1007 23:35:11.432980 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 07 23:35:20 crc kubenswrapper[4871]: I1007 23:35:20.982725 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:35:20 crc kubenswrapper[4871]: E1007 23:35:20.983652 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:35:34 crc kubenswrapper[4871]: I1007 23:35:34.983077 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:35:34 crc kubenswrapper[4871]: E1007 23:35:34.984056 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:35:47 crc kubenswrapper[4871]: I1007 23:35:47.982931 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:35:47 crc kubenswrapper[4871]: E1007 23:35:47.983959 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:35:59 crc kubenswrapper[4871]: I1007 23:35:59.982182 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:35:59 crc kubenswrapper[4871]: E1007 23:35:59.983164 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.519526 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 23:36:05 crc kubenswrapper[4871]: E1007 23:36:05.522698 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c37ba1e-a7b1-4398-8193-5396701840ea" containerName="mariadb-client" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.522867 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c37ba1e-a7b1-4398-8193-5396701840ea" containerName="mariadb-client" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.523155 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c37ba1e-a7b1-4398-8193-5396701840ea" containerName="mariadb-client" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.524108 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.526894 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.527896 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.530108 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-f9c6l" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.536885 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.538656 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.545257 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.551711 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.553565 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.568687 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.579852 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614606 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phd5h\" (UniqueName: \"kubernetes.io/projected/54438dbf-34ef-4626-acb1-8c0f4db20a88-kube-api-access-phd5h\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614660 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/258f0b30-db8c-4b1f-93e1-42df4430228c-config\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614706 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/54438dbf-34ef-4626-acb1-8c0f4db20a88-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614731 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0b39d035-a816-4208-8ead-2be6c63299d4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b39d035-a816-4208-8ead-2be6c63299d4\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614760 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/258f0b30-db8c-4b1f-93e1-42df4430228c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614836 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/258f0b30-db8c-4b1f-93e1-42df4430228c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614866 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54438dbf-34ef-4626-acb1-8c0f4db20a88-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614898 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/258f0b30-db8c-4b1f-93e1-42df4430228c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614921 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54438dbf-34ef-4626-acb1-8c0f4db20a88-config\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614943 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfjsd\" (UniqueName: \"kubernetes.io/projected/258f0b30-db8c-4b1f-93e1-42df4430228c-kube-api-access-lfjsd\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.614967 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54438dbf-34ef-4626-acb1-8c0f4db20a88-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.615079 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3bbbb072-4fb9-45d3-b37c-6e7c1655ca6b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bbbb072-4fb9-45d3-b37c-6e7c1655ca6b\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716187 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phd5h\" (UniqueName: \"kubernetes.io/projected/54438dbf-34ef-4626-acb1-8c0f4db20a88-kube-api-access-phd5h\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716240 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3b44f1fc-0f91-46fe-9c8b-5ea2b6a0d294\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b44f1fc-0f91-46fe-9c8b-5ea2b6a0d294\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716270 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/258f0b30-db8c-4b1f-93e1-42df4430228c-config\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716307 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/54438dbf-34ef-4626-acb1-8c0f4db20a88-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716327 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0b39d035-a816-4208-8ead-2be6c63299d4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b39d035-a816-4208-8ead-2be6c63299d4\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716349 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/258f0b30-db8c-4b1f-93e1-42df4430228c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716378 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbbjv\" (UniqueName: \"kubernetes.io/projected/eda861cc-b8ae-4a99-b651-a2f31601a006-kube-api-access-nbbjv\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716395 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/258f0b30-db8c-4b1f-93e1-42df4430228c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716412 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54438dbf-34ef-4626-acb1-8c0f4db20a88-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716475 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda861cc-b8ae-4a99-b651-a2f31601a006-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716517 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/258f0b30-db8c-4b1f-93e1-42df4430228c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716535 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54438dbf-34ef-4626-acb1-8c0f4db20a88-config\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716550 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eda861cc-b8ae-4a99-b651-a2f31601a006-config\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716570 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfjsd\" (UniqueName: \"kubernetes.io/projected/258f0b30-db8c-4b1f-93e1-42df4430228c-kube-api-access-lfjsd\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716592 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eda861cc-b8ae-4a99-b651-a2f31601a006-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716611 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54438dbf-34ef-4626-acb1-8c0f4db20a88-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716632 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3bbbb072-4fb9-45d3-b37c-6e7c1655ca6b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bbbb072-4fb9-45d3-b37c-6e7c1655ca6b\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.716659 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eda861cc-b8ae-4a99-b651-a2f31601a006-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.717156 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/258f0b30-db8c-4b1f-93e1-42df4430228c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.717919 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54438dbf-34ef-4626-acb1-8c0f4db20a88-config\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.718815 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54438dbf-34ef-4626-acb1-8c0f4db20a88-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.719486 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/54438dbf-34ef-4626-acb1-8c0f4db20a88-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.720569 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/258f0b30-db8c-4b1f-93e1-42df4430228c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.723332 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/258f0b30-db8c-4b1f-93e1-42df4430228c-config\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.724008 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54438dbf-34ef-4626-acb1-8c0f4db20a88-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.724121 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.724165 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0b39d035-a816-4208-8ead-2be6c63299d4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b39d035-a816-4208-8ead-2be6c63299d4\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a05587ee80fd5879cfd2ab613c44c84ba386f67b3e09888fa0d6d93cdc303bce/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.724470 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.724523 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3bbbb072-4fb9-45d3-b37c-6e7c1655ca6b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bbbb072-4fb9-45d3-b37c-6e7c1655ca6b\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5481ce226bc6935506bf5d6a47f45fc7c69986eec2325a0dfd216bb1f6eb03d9/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.730048 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/258f0b30-db8c-4b1f-93e1-42df4430228c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.738755 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phd5h\" (UniqueName: \"kubernetes.io/projected/54438dbf-34ef-4626-acb1-8c0f4db20a88-kube-api-access-phd5h\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.759054 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfjsd\" (UniqueName: \"kubernetes.io/projected/258f0b30-db8c-4b1f-93e1-42df4430228c-kube-api-access-lfjsd\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.760645 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.762394 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.765006 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.765313 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.769854 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-rxg2q" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.776756 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3bbbb072-4fb9-45d3-b37c-6e7c1655ca6b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bbbb072-4fb9-45d3-b37c-6e7c1655ca6b\") pod \"ovsdbserver-nb-0\" (UID: \"258f0b30-db8c-4b1f-93e1-42df4430228c\") " pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.777036 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.778895 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.783898 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.794244 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0b39d035-a816-4208-8ead-2be6c63299d4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b39d035-a816-4208-8ead-2be6c63299d4\") pod \"ovsdbserver-nb-1\" (UID: \"54438dbf-34ef-4626-acb1-8c0f4db20a88\") " pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.794574 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.795947 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.810421 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.816067 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.818405 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2bs2\" (UniqueName: \"kubernetes.io/projected/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-kube-api-access-l2bs2\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.818627 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.818764 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.818969 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-config\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821149 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3b44f1fc-0f91-46fe-9c8b-5ea2b6a0d294\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b44f1fc-0f91-46fe-9c8b-5ea2b6a0d294\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821190 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821231 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2282fef5-1de5-458e-8f94-2bd8d8f6b44b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2282fef5-1de5-458e-8f94-2bd8d8f6b44b\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821255 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821279 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821327 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbbjv\" (UniqueName: \"kubernetes.io/projected/eda861cc-b8ae-4a99-b651-a2f31601a006-kube-api-access-nbbjv\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821366 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5e7c21de-5eda-4c7e-9243-b1127f1f70a7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5e7c21de-5eda-4c7e-9243-b1127f1f70a7\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821393 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda861cc-b8ae-4a99-b651-a2f31601a006-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821702 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821835 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-config\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821869 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eda861cc-b8ae-4a99-b651-a2f31601a006-config\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.821934 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eda861cc-b8ae-4a99-b651-a2f31601a006-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.823158 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eda861cc-b8ae-4a99-b651-a2f31601a006-config\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.823330 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eda861cc-b8ae-4a99-b651-a2f31601a006-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.823697 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eda861cc-b8ae-4a99-b651-a2f31601a006-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.823364 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh8xx\" (UniqueName: \"kubernetes.io/projected/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-kube-api-access-zh8xx\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.825297 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eda861cc-b8ae-4a99-b651-a2f31601a006-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.827764 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.827824 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3b44f1fc-0f91-46fe-9c8b-5ea2b6a0d294\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b44f1fc-0f91-46fe-9c8b-5ea2b6a0d294\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/97b0045adcfac88757489d252dd81fa0dae2971edac379a9dda9de1888fe86f4/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.832052 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda861cc-b8ae-4a99-b651-a2f31601a006-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.841367 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbbjv\" (UniqueName: \"kubernetes.io/projected/eda861cc-b8ae-4a99-b651-a2f31601a006-kube-api-access-nbbjv\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.854088 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3b44f1fc-0f91-46fe-9c8b-5ea2b6a0d294\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b44f1fc-0f91-46fe-9c8b-5ea2b6a0d294\") pod \"ovsdbserver-nb-2\" (UID: \"eda861cc-b8ae-4a99-b651-a2f31601a006\") " pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.872156 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.882168 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.898372 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925111 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2282fef5-1de5-458e-8f94-2bd8d8f6b44b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2282fef5-1de5-458e-8f94-2bd8d8f6b44b\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925180 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925209 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925268 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5e7c21de-5eda-4c7e-9243-b1127f1f70a7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5e7c21de-5eda-4c7e-9243-b1127f1f70a7\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925303 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925337 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fc0b5358-5cd6-4c83-b624-fb6418b7cc16\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fc0b5358-5cd6-4c83-b624-fb6418b7cc16\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925364 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-config\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925392 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d38b4227-32df-4361-b559-bd3c7ef7e741-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925424 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9gfr\" (UniqueName: \"kubernetes.io/projected/d38b4227-32df-4361-b559-bd3c7ef7e741-kube-api-access-j9gfr\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925449 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d38b4227-32df-4361-b559-bd3c7ef7e741-config\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925491 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh8xx\" (UniqueName: \"kubernetes.io/projected/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-kube-api-access-zh8xx\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925518 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38b4227-32df-4361-b559-bd3c7ef7e741-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925544 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2bs2\" (UniqueName: \"kubernetes.io/projected/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-kube-api-access-l2bs2\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925572 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925596 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d38b4227-32df-4361-b559-bd3c7ef7e741-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925638 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925664 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-config\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.925716 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.927361 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.927407 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-config\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.927412 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.927686 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.928709 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-config\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.930822 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.930867 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5e7c21de-5eda-4c7e-9243-b1127f1f70a7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5e7c21de-5eda-4c7e-9243-b1127f1f70a7\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/92ea33edd478527c3e43e093eb1fd47bd843f366e705516863dcdbbe5b693ca3/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.933613 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.934898 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.939946 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.939995 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2282fef5-1de5-458e-8f94-2bd8d8f6b44b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2282fef5-1de5-458e-8f94-2bd8d8f6b44b\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c4ea324767efccef86699dbf75ea09fd46ec7c22e083d6581315fc8729ed7e52/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.944864 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2bs2\" (UniqueName: \"kubernetes.io/projected/f2d065bc-0f6a-4e23-8ffe-b9d07d512d08-kube-api-access-l2bs2\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.945319 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.945862 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh8xx\" (UniqueName: \"kubernetes.io/projected/faf833c3-7d4d-4c64-b466-3c48ef1a42c9-kube-api-access-zh8xx\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:05 crc kubenswrapper[4871]: I1007 23:36:05.968526 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5e7c21de-5eda-4c7e-9243-b1127f1f70a7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5e7c21de-5eda-4c7e-9243-b1127f1f70a7\") pod \"ovsdbserver-sb-2\" (UID: \"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08\") " pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.015508 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2282fef5-1de5-458e-8f94-2bd8d8f6b44b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2282fef5-1de5-458e-8f94-2bd8d8f6b44b\") pod \"ovsdbserver-sb-0\" (UID: \"faf833c3-7d4d-4c64-b466-3c48ef1a42c9\") " pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.027257 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fc0b5358-5cd6-4c83-b624-fb6418b7cc16\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fc0b5358-5cd6-4c83-b624-fb6418b7cc16\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.027310 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d38b4227-32df-4361-b559-bd3c7ef7e741-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.027341 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9gfr\" (UniqueName: \"kubernetes.io/projected/d38b4227-32df-4361-b559-bd3c7ef7e741-kube-api-access-j9gfr\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.027361 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d38b4227-32df-4361-b559-bd3c7ef7e741-config\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.027410 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38b4227-32df-4361-b559-bd3c7ef7e741-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.027435 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d38b4227-32df-4361-b559-bd3c7ef7e741-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.028682 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d38b4227-32df-4361-b559-bd3c7ef7e741-config\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.029818 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d38b4227-32df-4361-b559-bd3c7ef7e741-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.030608 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d38b4227-32df-4361-b559-bd3c7ef7e741-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.033352 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.033388 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fc0b5358-5cd6-4c83-b624-fb6418b7cc16\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fc0b5358-5cd6-4c83-b624-fb6418b7cc16\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d894bcc1bbcaab8a532bc5c1af6f6cd2afb08a7651c448ccdf32900c61fd69df/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.042508 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38b4227-32df-4361-b559-bd3c7ef7e741-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.046978 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9gfr\" (UniqueName: \"kubernetes.io/projected/d38b4227-32df-4361-b559-bd3c7ef7e741-kube-api-access-j9gfr\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.061635 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fc0b5358-5cd6-4c83-b624-fb6418b7cc16\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fc0b5358-5cd6-4c83-b624-fb6418b7cc16\") pod \"ovsdbserver-sb-1\" (UID: \"d38b4227-32df-4361-b559-bd3c7ef7e741\") " pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.248721 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.258941 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.266261 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.450962 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.556547 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 07 23:36:06 crc kubenswrapper[4871]: W1007 23:36:06.565472 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeda861cc_b8ae_4a99_b651_a2f31601a006.slice/crio-386f9b6f747900a3d06833be30c2a3e62f734fd30969d61db4275a8925f24275 WatchSource:0}: Error finding container 386f9b6f747900a3d06833be30c2a3e62f734fd30969d61db4275a8925f24275: Status 404 returned error can't find the container with id 386f9b6f747900a3d06833be30c2a3e62f734fd30969d61db4275a8925f24275 Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.826764 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.917637 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 23:36:06 crc kubenswrapper[4871]: W1007 23:36:06.925599 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaf833c3_7d4d_4c64_b466_3c48ef1a42c9.slice/crio-24a0b8169063fff42343077a14d3a488fff715dbf06c5f5d4754f8aed2868063 WatchSource:0}: Error finding container 24a0b8169063fff42343077a14d3a488fff715dbf06c5f5d4754f8aed2868063: Status 404 returned error can't find the container with id 24a0b8169063fff42343077a14d3a488fff715dbf06c5f5d4754f8aed2868063 Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.945183 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"eda861cc-b8ae-4a99-b651-a2f31601a006","Type":"ContainerStarted","Data":"478e147e515074289d4db2a7e5bc67aebd83297aefd1606c6877e53b3ee00dc2"} Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.945269 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"eda861cc-b8ae-4a99-b651-a2f31601a006","Type":"ContainerStarted","Data":"386f9b6f747900a3d06833be30c2a3e62f734fd30969d61db4275a8925f24275"} Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.948431 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"258f0b30-db8c-4b1f-93e1-42df4430228c","Type":"ContainerStarted","Data":"41d9c4b9f28546f5c83b50efd8ed0b349de54b53b79887c7acc16a675bec1462"} Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.948486 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"258f0b30-db8c-4b1f-93e1-42df4430228c","Type":"ContainerStarted","Data":"5536ec2329fb56a7698dbf1875c36140838ae49f984e1487637199bdc6852bfe"} Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.948498 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"258f0b30-db8c-4b1f-93e1-42df4430228c","Type":"ContainerStarted","Data":"a59b462aef51a21fd03b925afbffe122446b0c46abd4f9a6660c2d76b7f5cca7"} Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.951671 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"d38b4227-32df-4361-b559-bd3c7ef7e741","Type":"ContainerStarted","Data":"ec2a99725fd479e872cb7403f7e641a9b1e736174a95ec92ea2e0aeb0d253ca7"} Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.952694 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"faf833c3-7d4d-4c64-b466-3c48ef1a42c9","Type":"ContainerStarted","Data":"24a0b8169063fff42343077a14d3a488fff715dbf06c5f5d4754f8aed2868063"} Oct 07 23:36:06 crc kubenswrapper[4871]: I1007 23:36:06.971322 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=2.971304897 podStartE2EDuration="2.971304897s" podCreationTimestamp="2025-10-07 23:36:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:06.967080005 +0000 UTC m=+5240.769778078" watchObservedRunningTime="2025-10-07 23:36:06.971304897 +0000 UTC m=+5240.774002970" Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.164900 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 07 23:36:07 crc kubenswrapper[4871]: W1007 23:36:07.169422 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54438dbf_34ef_4626_acb1_8c0f4db20a88.slice/crio-a75a2a0af4b959c6dd49ef348ee5b2e2ea505eec5e46ff2d695cc4b7c5f3cb93 WatchSource:0}: Error finding container a75a2a0af4b959c6dd49ef348ee5b2e2ea505eec5e46ff2d695cc4b7c5f3cb93: Status 404 returned error can't find the container with id a75a2a0af4b959c6dd49ef348ee5b2e2ea505eec5e46ff2d695cc4b7c5f3cb93 Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.806003 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 07 23:36:07 crc kubenswrapper[4871]: W1007 23:36:07.806781 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2d065bc_0f6a_4e23_8ffe_b9d07d512d08.slice/crio-54a446a27d6a050413f355d3eea3acf856d2baf471c52f592ce17f1233a887d8 WatchSource:0}: Error finding container 54a446a27d6a050413f355d3eea3acf856d2baf471c52f592ce17f1233a887d8: Status 404 returned error can't find the container with id 54a446a27d6a050413f355d3eea3acf856d2baf471c52f592ce17f1233a887d8 Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.965176 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08","Type":"ContainerStarted","Data":"54a446a27d6a050413f355d3eea3acf856d2baf471c52f592ce17f1233a887d8"} Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.968350 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"faf833c3-7d4d-4c64-b466-3c48ef1a42c9","Type":"ContainerStarted","Data":"24a5f761038f6528d1a1e18e7b657bb6710ab60882dc02124bb2d0fbdae969ff"} Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.968399 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"faf833c3-7d4d-4c64-b466-3c48ef1a42c9","Type":"ContainerStarted","Data":"0303275c88004a921583b18b64b52de12cfaf5a5121a3d2487113f97a5a2bf35"} Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.973257 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"54438dbf-34ef-4626-acb1-8c0f4db20a88","Type":"ContainerStarted","Data":"58d65c058bfea0144ac3968a5e4725c9f5d27b6f81a4111999e1f444483ddccc"} Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.973291 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"54438dbf-34ef-4626-acb1-8c0f4db20a88","Type":"ContainerStarted","Data":"315e7700cbd93782a394b21f067201c7e7b4d274a843a048900cf58f34a44374"} Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.973307 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"54438dbf-34ef-4626-acb1-8c0f4db20a88","Type":"ContainerStarted","Data":"a75a2a0af4b959c6dd49ef348ee5b2e2ea505eec5e46ff2d695cc4b7c5f3cb93"} Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.977610 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"eda861cc-b8ae-4a99-b651-a2f31601a006","Type":"ContainerStarted","Data":"9b5106a7dbccec7ef266f7a259e523cd21b379587b2871179ffb83b0175135d9"} Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.980607 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"d38b4227-32df-4361-b559-bd3c7ef7e741","Type":"ContainerStarted","Data":"cd10e5841a19642544fb2d933da8f1a83c994f85d4b876969b13f62b4b16b37b"} Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.980726 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"d38b4227-32df-4361-b559-bd3c7ef7e741","Type":"ContainerStarted","Data":"ba6dfd675b1197fbc7e4036a7a2673c986800bc239c562c8d71eb98c559d5fbd"} Oct 07 23:36:07 crc kubenswrapper[4871]: I1007 23:36:07.996378 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.9963570969999997 podStartE2EDuration="3.996357097s" podCreationTimestamp="2025-10-07 23:36:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:07.993859761 +0000 UTC m=+5241.796557874" watchObservedRunningTime="2025-10-07 23:36:07.996357097 +0000 UTC m=+5241.799055170" Oct 07 23:36:08 crc kubenswrapper[4871]: I1007 23:36:08.016114 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.016097868 podStartE2EDuration="4.016097868s" podCreationTimestamp="2025-10-07 23:36:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:08.010940352 +0000 UTC m=+5241.813638425" watchObservedRunningTime="2025-10-07 23:36:08.016097868 +0000 UTC m=+5241.818795941" Oct 07 23:36:08 crc kubenswrapper[4871]: I1007 23:36:08.038170 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=4.038120129 podStartE2EDuration="4.038120129s" podCreationTimestamp="2025-10-07 23:36:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:08.035274004 +0000 UTC m=+5241.837972097" watchObservedRunningTime="2025-10-07 23:36:08.038120129 +0000 UTC m=+5241.840818212" Oct 07 23:36:08 crc kubenswrapper[4871]: I1007 23:36:08.060407 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.060385607 podStartE2EDuration="4.060385607s" podCreationTimestamp="2025-10-07 23:36:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:08.059694799 +0000 UTC m=+5241.862392892" watchObservedRunningTime="2025-10-07 23:36:08.060385607 +0000 UTC m=+5241.863083690" Oct 07 23:36:08 crc kubenswrapper[4871]: I1007 23:36:08.872726 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:08 crc kubenswrapper[4871]: I1007 23:36:08.882519 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:08 crc kubenswrapper[4871]: I1007 23:36:08.898893 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:09 crc kubenswrapper[4871]: I1007 23:36:09.027483 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08","Type":"ContainerStarted","Data":"8121846e3fff41717c1ba0e9f747750b9863788c78457f6e328557e2f52d562d"} Oct 07 23:36:09 crc kubenswrapper[4871]: I1007 23:36:09.027556 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"f2d065bc-0f6a-4e23-8ffe-b9d07d512d08","Type":"ContainerStarted","Data":"4f0a35f234842058238cc929d33bbf88b1e9e070492f8d36c2efcf67c637f154"} Oct 07 23:36:09 crc kubenswrapper[4871]: I1007 23:36:09.050072 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=5.050054494 podStartE2EDuration="5.050054494s" podCreationTimestamp="2025-10-07 23:36:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:09.044573759 +0000 UTC m=+5242.847271872" watchObservedRunningTime="2025-10-07 23:36:09.050054494 +0000 UTC m=+5242.852752567" Oct 07 23:36:09 crc kubenswrapper[4871]: I1007 23:36:09.249737 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:09 crc kubenswrapper[4871]: I1007 23:36:09.259492 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:09 crc kubenswrapper[4871]: I1007 23:36:09.267077 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:10 crc kubenswrapper[4871]: I1007 23:36:10.872844 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:10 crc kubenswrapper[4871]: I1007 23:36:10.883153 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:10 crc kubenswrapper[4871]: I1007 23:36:10.899359 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:11 crc kubenswrapper[4871]: I1007 23:36:11.250386 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:11 crc kubenswrapper[4871]: I1007 23:36:11.259931 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:11 crc kubenswrapper[4871]: I1007 23:36:11.267052 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:11 crc kubenswrapper[4871]: I1007 23:36:11.945402 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:11 crc kubenswrapper[4871]: I1007 23:36:11.958573 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:11 crc kubenswrapper[4871]: I1007 23:36:11.962430 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:11 crc kubenswrapper[4871]: I1007 23:36:11.982954 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:36:11 crc kubenswrapper[4871]: E1007 23:36:11.983363 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.020944 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.098021 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.104251 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.245642 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c45d5695-j76jp"] Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.246853 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.250762 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.271376 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c45d5695-j76jp"] Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.298929 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.318712 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.325126 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.339767 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-dns-svc\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.339851 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-config\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.339912 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-ovsdbserver-nb\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.339937 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwffk\" (UniqueName: \"kubernetes.io/projected/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-kube-api-access-bwffk\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.344379 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.372310 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.442341 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-ovsdbserver-nb\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.442394 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwffk\" (UniqueName: \"kubernetes.io/projected/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-kube-api-access-bwffk\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.442496 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-dns-svc\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.442573 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-config\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.443803 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-dns-svc\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.443786 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-ovsdbserver-nb\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.443923 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-config\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.464294 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwffk\" (UniqueName: \"kubernetes.io/projected/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-kube-api-access-bwffk\") pod \"dnsmasq-dns-7c45d5695-j76jp\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.566374 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.787885 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c45d5695-j76jp"] Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.823975 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69765ddff5-59x29"] Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.825199 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.829088 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.837160 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69765ddff5-59x29"] Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.856894 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7k24\" (UniqueName: \"kubernetes.io/projected/501ccea2-62f3-426b-91cd-3d1423944ef0-kube-api-access-r7k24\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.856953 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-sb\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.857087 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-nb\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.857137 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-config\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.857186 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-dns-svc\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.959017 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-nb\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.959094 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-config\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.959118 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-dns-svc\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.959174 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7k24\" (UniqueName: \"kubernetes.io/projected/501ccea2-62f3-426b-91cd-3d1423944ef0-kube-api-access-r7k24\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.959199 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-sb\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.960081 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-config\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.960119 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-sb\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.960428 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-dns-svc\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.960766 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-nb\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:12 crc kubenswrapper[4871]: I1007 23:36:12.995237 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7k24\" (UniqueName: \"kubernetes.io/projected/501ccea2-62f3-426b-91cd-3d1423944ef0-kube-api-access-r7k24\") pod \"dnsmasq-dns-69765ddff5-59x29\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:13 crc kubenswrapper[4871]: I1007 23:36:13.126515 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c45d5695-j76jp"] Oct 07 23:36:13 crc kubenswrapper[4871]: W1007 23:36:13.127115 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod824e2ab2_5af7_4885_b8ba_c7ac0c8311ea.slice/crio-2dc21d29520a7d27cc2ffd4dbb4e715954c6590328d260c97803ad53b9d4708f WatchSource:0}: Error finding container 2dc21d29520a7d27cc2ffd4dbb4e715954c6590328d260c97803ad53b9d4708f: Status 404 returned error can't find the container with id 2dc21d29520a7d27cc2ffd4dbb4e715954c6590328d260c97803ad53b9d4708f Oct 07 23:36:13 crc kubenswrapper[4871]: I1007 23:36:13.128236 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 07 23:36:13 crc kubenswrapper[4871]: I1007 23:36:13.151953 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:13 crc kubenswrapper[4871]: I1007 23:36:13.628846 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69765ddff5-59x29"] Oct 07 23:36:13 crc kubenswrapper[4871]: W1007 23:36:13.633014 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod501ccea2_62f3_426b_91cd_3d1423944ef0.slice/crio-7a66befe648211bf82eeb55fefb4d782fe30bbec48c1f13aefd909964c196da3 WatchSource:0}: Error finding container 7a66befe648211bf82eeb55fefb4d782fe30bbec48c1f13aefd909964c196da3: Status 404 returned error can't find the container with id 7a66befe648211bf82eeb55fefb4d782fe30bbec48c1f13aefd909964c196da3 Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.071820 4871 generic.go:334] "Generic (PLEG): container finished" podID="501ccea2-62f3-426b-91cd-3d1423944ef0" containerID="4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e" exitCode=0 Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.071879 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69765ddff5-59x29" event={"ID":"501ccea2-62f3-426b-91cd-3d1423944ef0","Type":"ContainerDied","Data":"4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e"} Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.072245 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69765ddff5-59x29" event={"ID":"501ccea2-62f3-426b-91cd-3d1423944ef0","Type":"ContainerStarted","Data":"7a66befe648211bf82eeb55fefb4d782fe30bbec48c1f13aefd909964c196da3"} Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.074127 4871 generic.go:334] "Generic (PLEG): container finished" podID="824e2ab2-5af7-4885-b8ba-c7ac0c8311ea" containerID="fb9b74a8a9e7df30aae7489ca728c34d20d89985460a6195e8b9c79ce3c02fc4" exitCode=0 Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.074224 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c45d5695-j76jp" event={"ID":"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea","Type":"ContainerDied","Data":"fb9b74a8a9e7df30aae7489ca728c34d20d89985460a6195e8b9c79ce3c02fc4"} Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.074246 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c45d5695-j76jp" event={"ID":"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea","Type":"ContainerStarted","Data":"2dc21d29520a7d27cc2ffd4dbb4e715954c6590328d260c97803ad53b9d4708f"} Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.387019 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.514449 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-config\") pod \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.514531 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-dns-svc\") pod \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.514575 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-ovsdbserver-nb\") pod \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.514592 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwffk\" (UniqueName: \"kubernetes.io/projected/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-kube-api-access-bwffk\") pod \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\" (UID: \"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea\") " Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.518863 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-kube-api-access-bwffk" (OuterVolumeSpecName: "kube-api-access-bwffk") pod "824e2ab2-5af7-4885-b8ba-c7ac0c8311ea" (UID: "824e2ab2-5af7-4885-b8ba-c7ac0c8311ea"). InnerVolumeSpecName "kube-api-access-bwffk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.534517 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-config" (OuterVolumeSpecName: "config") pod "824e2ab2-5af7-4885-b8ba-c7ac0c8311ea" (UID: "824e2ab2-5af7-4885-b8ba-c7ac0c8311ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.536599 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "824e2ab2-5af7-4885-b8ba-c7ac0c8311ea" (UID: "824e2ab2-5af7-4885-b8ba-c7ac0c8311ea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.540273 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "824e2ab2-5af7-4885-b8ba-c7ac0c8311ea" (UID: "824e2ab2-5af7-4885-b8ba-c7ac0c8311ea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.616355 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.616388 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.616398 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:14 crc kubenswrapper[4871]: I1007 23:36:14.616412 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwffk\" (UniqueName: \"kubernetes.io/projected/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea-kube-api-access-bwffk\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.093392 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c45d5695-j76jp" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.093391 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c45d5695-j76jp" event={"ID":"824e2ab2-5af7-4885-b8ba-c7ac0c8311ea","Type":"ContainerDied","Data":"2dc21d29520a7d27cc2ffd4dbb4e715954c6590328d260c97803ad53b9d4708f"} Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.094237 4871 scope.go:117] "RemoveContainer" containerID="fb9b74a8a9e7df30aae7489ca728c34d20d89985460a6195e8b9c79ce3c02fc4" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.097772 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69765ddff5-59x29" event={"ID":"501ccea2-62f3-426b-91cd-3d1423944ef0","Type":"ContainerStarted","Data":"d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d"} Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.098099 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.136367 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69765ddff5-59x29" podStartSLOduration=3.136351786 podStartE2EDuration="3.136351786s" podCreationTimestamp="2025-10-07 23:36:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:15.13155603 +0000 UTC m=+5248.934254123" watchObservedRunningTime="2025-10-07 23:36:15.136351786 +0000 UTC m=+5248.939049859" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.197816 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c45d5695-j76jp"] Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.203823 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c45d5695-j76jp"] Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.460636 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 07 23:36:15 crc kubenswrapper[4871]: E1007 23:36:15.461224 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824e2ab2-5af7-4885-b8ba-c7ac0c8311ea" containerName="init" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.461256 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="824e2ab2-5af7-4885-b8ba-c7ac0c8311ea" containerName="init" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.461561 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="824e2ab2-5af7-4885-b8ba-c7ac0c8311ea" containerName="init" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.462485 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.464356 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.480277 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.635825 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw4t7\" (UniqueName: \"kubernetes.io/projected/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-kube-api-access-qw4t7\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.635879 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.636114 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.738666 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw4t7\" (UniqueName: \"kubernetes.io/projected/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-kube-api-access-qw4t7\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.738830 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.738950 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.741658 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.741689 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7ad648429b98d91158b4d56d74b91dabe918470be1b346e2df9eb5b874ca090e/globalmount\"" pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.747642 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.756537 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw4t7\" (UniqueName: \"kubernetes.io/projected/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-kube-api-access-qw4t7\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.770371 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\") pod \"ovn-copy-data\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " pod="openstack/ovn-copy-data" Oct 07 23:36:15 crc kubenswrapper[4871]: I1007 23:36:15.786448 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 07 23:36:16 crc kubenswrapper[4871]: I1007 23:36:16.141506 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 07 23:36:16 crc kubenswrapper[4871]: W1007 23:36:16.146897 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6206f2_ab54_4ec5_97e6_ed2460f6ee30.slice/crio-bfb19b21f38558b0a6e163e2a8be829a778c9f7dfc9ec71f2353acc99e1d8f42 WatchSource:0}: Error finding container bfb19b21f38558b0a6e163e2a8be829a778c9f7dfc9ec71f2353acc99e1d8f42: Status 404 returned error can't find the container with id bfb19b21f38558b0a6e163e2a8be829a778c9f7dfc9ec71f2353acc99e1d8f42 Oct 07 23:36:16 crc kubenswrapper[4871]: I1007 23:36:16.149737 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 23:36:17 crc kubenswrapper[4871]: I1007 23:36:17.001981 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="824e2ab2-5af7-4885-b8ba-c7ac0c8311ea" path="/var/lib/kubelet/pods/824e2ab2-5af7-4885-b8ba-c7ac0c8311ea/volumes" Oct 07 23:36:17 crc kubenswrapper[4871]: I1007 23:36:17.122830 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30","Type":"ContainerStarted","Data":"76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455"} Oct 07 23:36:17 crc kubenswrapper[4871]: I1007 23:36:17.122892 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30","Type":"ContainerStarted","Data":"bfb19b21f38558b0a6e163e2a8be829a778c9f7dfc9ec71f2353acc99e1d8f42"} Oct 07 23:36:17 crc kubenswrapper[4871]: I1007 23:36:17.147202 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.573914747 podStartE2EDuration="3.147176131s" podCreationTimestamp="2025-10-07 23:36:14 +0000 UTC" firstStartedPulling="2025-10-07 23:36:16.149481122 +0000 UTC m=+5249.952179195" lastFinishedPulling="2025-10-07 23:36:16.722742506 +0000 UTC m=+5250.525440579" observedRunningTime="2025-10-07 23:36:17.140594127 +0000 UTC m=+5250.943292260" watchObservedRunningTime="2025-10-07 23:36:17.147176131 +0000 UTC m=+5250.949874234" Oct 07 23:36:20 crc kubenswrapper[4871]: E1007 23:36:20.389025 4871 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.68:45686->38.102.83.68:41109: read tcp 38.102.83.68:45686->38.102.83.68:41109: read: connection reset by peer Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.624876 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.626664 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.629047 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-pt84n" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.629914 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.629970 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.635657 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.665765 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.665828 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-config\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.665862 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.665910 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flhc6\" (UniqueName: \"kubernetes.io/projected/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-kube-api-access-flhc6\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.665946 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-scripts\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.767703 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.767748 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-config\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.767784 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.767903 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flhc6\" (UniqueName: \"kubernetes.io/projected/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-kube-api-access-flhc6\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.767933 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-scripts\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.768581 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.768733 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-scripts\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.768732 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-config\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.776551 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.786518 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flhc6\" (UniqueName: \"kubernetes.io/projected/df2f59e7-06b2-42a4-bd21-d972ad2f89d5-kube-api-access-flhc6\") pod \"ovn-northd-0\" (UID: \"df2f59e7-06b2-42a4-bd21-d972ad2f89d5\") " pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.982731 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 23:36:22 crc kubenswrapper[4871]: I1007 23:36:22.983174 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:36:22 crc kubenswrapper[4871]: E1007 23:36:22.983683 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.155558 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.215717 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-lbns8"] Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.215981 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" podUID="5532d18c-3b32-4d80-9db3-1e20e40f4c34" containerName="dnsmasq-dns" containerID="cri-o://299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9" gracePeriod=10 Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.496189 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.644752 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.682940 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwmv8\" (UniqueName: \"kubernetes.io/projected/5532d18c-3b32-4d80-9db3-1e20e40f4c34-kube-api-access-kwmv8\") pod \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.682997 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-config\") pod \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.683103 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-dns-svc\") pod \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\" (UID: \"5532d18c-3b32-4d80-9db3-1e20e40f4c34\") " Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.689726 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5532d18c-3b32-4d80-9db3-1e20e40f4c34-kube-api-access-kwmv8" (OuterVolumeSpecName: "kube-api-access-kwmv8") pod "5532d18c-3b32-4d80-9db3-1e20e40f4c34" (UID: "5532d18c-3b32-4d80-9db3-1e20e40f4c34"). InnerVolumeSpecName "kube-api-access-kwmv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.724406 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5532d18c-3b32-4d80-9db3-1e20e40f4c34" (UID: "5532d18c-3b32-4d80-9db3-1e20e40f4c34"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.730085 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-config" (OuterVolumeSpecName: "config") pod "5532d18c-3b32-4d80-9db3-1e20e40f4c34" (UID: "5532d18c-3b32-4d80-9db3-1e20e40f4c34"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.784628 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.784663 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwmv8\" (UniqueName: \"kubernetes.io/projected/5532d18c-3b32-4d80-9db3-1e20e40f4c34-kube-api-access-kwmv8\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:23 crc kubenswrapper[4871]: I1007 23:36:23.784679 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5532d18c-3b32-4d80-9db3-1e20e40f4c34-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.203124 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"df2f59e7-06b2-42a4-bd21-d972ad2f89d5","Type":"ContainerStarted","Data":"c10417ad2f5f34d7195fac58f1790bea281f14fcca5bda9344176bbdac1ed27e"} Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.203170 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"df2f59e7-06b2-42a4-bd21-d972ad2f89d5","Type":"ContainerStarted","Data":"6cce3478c6beae1243295fbf9a6cff3d8a7a33d55c505cbf02a3045e2e791e6d"} Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.203183 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"df2f59e7-06b2-42a4-bd21-d972ad2f89d5","Type":"ContainerStarted","Data":"326ea09204b9083faa2c12b66e5245a35c0f260cb0875f310122d57fbd24169a"} Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.203375 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.207495 4871 generic.go:334] "Generic (PLEG): container finished" podID="5532d18c-3b32-4d80-9db3-1e20e40f4c34" containerID="299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9" exitCode=0 Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.207563 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" event={"ID":"5532d18c-3b32-4d80-9db3-1e20e40f4c34","Type":"ContainerDied","Data":"299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9"} Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.207600 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" event={"ID":"5532d18c-3b32-4d80-9db3-1e20e40f4c34","Type":"ContainerDied","Data":"6a556f5d5b3270783ac818fb5c7b1a6d78feff8e1821398c06ad2d7487524563"} Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.207637 4871 scope.go:117] "RemoveContainer" containerID="299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.207778 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fdc957c47-lbns8" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.233401 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.233376822 podStartE2EDuration="2.233376822s" podCreationTimestamp="2025-10-07 23:36:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:24.224489817 +0000 UTC m=+5258.027187900" watchObservedRunningTime="2025-10-07 23:36:24.233376822 +0000 UTC m=+5258.036074905" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.247613 4871 scope.go:117] "RemoveContainer" containerID="f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.272690 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-lbns8"] Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.279566 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-lbns8"] Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.291198 4871 scope.go:117] "RemoveContainer" containerID="299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9" Oct 07 23:36:24 crc kubenswrapper[4871]: E1007 23:36:24.292572 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9\": container with ID starting with 299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9 not found: ID does not exist" containerID="299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.292643 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9"} err="failed to get container status \"299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9\": rpc error: code = NotFound desc = could not find container \"299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9\": container with ID starting with 299d26fc8f1a5c8341d6f3e00995dcd20c3cab007bcd20052ba077194622e3c9 not found: ID does not exist" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.292696 4871 scope.go:117] "RemoveContainer" containerID="f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8" Oct 07 23:36:24 crc kubenswrapper[4871]: E1007 23:36:24.293165 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8\": container with ID starting with f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8 not found: ID does not exist" containerID="f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.293233 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8"} err="failed to get container status \"f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8\": rpc error: code = NotFound desc = could not find container \"f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8\": container with ID starting with f0772f34f21a13c2115041d50d47223e963d625e39f1dcbcdb9d3bb85cd2c8c8 not found: ID does not exist" Oct 07 23:36:24 crc kubenswrapper[4871]: I1007 23:36:24.997150 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5532d18c-3b32-4d80-9db3-1e20e40f4c34" path="/var/lib/kubelet/pods/5532d18c-3b32-4d80-9db3-1e20e40f4c34/volumes" Oct 07 23:36:27 crc kubenswrapper[4871]: I1007 23:36:27.743293 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-mldkt"] Oct 07 23:36:27 crc kubenswrapper[4871]: E1007 23:36:27.743808 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5532d18c-3b32-4d80-9db3-1e20e40f4c34" containerName="init" Oct 07 23:36:27 crc kubenswrapper[4871]: I1007 23:36:27.743820 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5532d18c-3b32-4d80-9db3-1e20e40f4c34" containerName="init" Oct 07 23:36:27 crc kubenswrapper[4871]: E1007 23:36:27.743838 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5532d18c-3b32-4d80-9db3-1e20e40f4c34" containerName="dnsmasq-dns" Oct 07 23:36:27 crc kubenswrapper[4871]: I1007 23:36:27.743845 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5532d18c-3b32-4d80-9db3-1e20e40f4c34" containerName="dnsmasq-dns" Oct 07 23:36:27 crc kubenswrapper[4871]: I1007 23:36:27.744020 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5532d18c-3b32-4d80-9db3-1e20e40f4c34" containerName="dnsmasq-dns" Oct 07 23:36:27 crc kubenswrapper[4871]: I1007 23:36:27.744513 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mldkt" Oct 07 23:36:27 crc kubenswrapper[4871]: I1007 23:36:27.751109 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mldkt"] Oct 07 23:36:27 crc kubenswrapper[4871]: I1007 23:36:27.876824 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfgft\" (UniqueName: \"kubernetes.io/projected/9936e5dd-19e9-4fea-89a2-f4c0f202f880-kube-api-access-rfgft\") pod \"keystone-db-create-mldkt\" (UID: \"9936e5dd-19e9-4fea-89a2-f4c0f202f880\") " pod="openstack/keystone-db-create-mldkt" Oct 07 23:36:27 crc kubenswrapper[4871]: I1007 23:36:27.978450 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfgft\" (UniqueName: \"kubernetes.io/projected/9936e5dd-19e9-4fea-89a2-f4c0f202f880-kube-api-access-rfgft\") pod \"keystone-db-create-mldkt\" (UID: \"9936e5dd-19e9-4fea-89a2-f4c0f202f880\") " pod="openstack/keystone-db-create-mldkt" Oct 07 23:36:28 crc kubenswrapper[4871]: I1007 23:36:28.020018 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfgft\" (UniqueName: \"kubernetes.io/projected/9936e5dd-19e9-4fea-89a2-f4c0f202f880-kube-api-access-rfgft\") pod \"keystone-db-create-mldkt\" (UID: \"9936e5dd-19e9-4fea-89a2-f4c0f202f880\") " pod="openstack/keystone-db-create-mldkt" Oct 07 23:36:28 crc kubenswrapper[4871]: I1007 23:36:28.071642 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mldkt" Oct 07 23:36:28 crc kubenswrapper[4871]: I1007 23:36:28.517779 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mldkt"] Oct 07 23:36:29 crc kubenswrapper[4871]: I1007 23:36:29.270146 4871 generic.go:334] "Generic (PLEG): container finished" podID="9936e5dd-19e9-4fea-89a2-f4c0f202f880" containerID="ace0d887e0e6ec937663f092e2b65b870376917e90f9053593a716352bc8f229" exitCode=0 Oct 07 23:36:29 crc kubenswrapper[4871]: I1007 23:36:29.270224 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mldkt" event={"ID":"9936e5dd-19e9-4fea-89a2-f4c0f202f880","Type":"ContainerDied","Data":"ace0d887e0e6ec937663f092e2b65b870376917e90f9053593a716352bc8f229"} Oct 07 23:36:29 crc kubenswrapper[4871]: I1007 23:36:29.270577 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mldkt" event={"ID":"9936e5dd-19e9-4fea-89a2-f4c0f202f880","Type":"ContainerStarted","Data":"5ad0558a040d2f1ae1622edbee0dd430b3234e259933a1446d1329363b2eb79c"} Oct 07 23:36:30 crc kubenswrapper[4871]: I1007 23:36:30.653343 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mldkt" Oct 07 23:36:30 crc kubenswrapper[4871]: I1007 23:36:30.727350 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfgft\" (UniqueName: \"kubernetes.io/projected/9936e5dd-19e9-4fea-89a2-f4c0f202f880-kube-api-access-rfgft\") pod \"9936e5dd-19e9-4fea-89a2-f4c0f202f880\" (UID: \"9936e5dd-19e9-4fea-89a2-f4c0f202f880\") " Oct 07 23:36:30 crc kubenswrapper[4871]: I1007 23:36:30.733678 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9936e5dd-19e9-4fea-89a2-f4c0f202f880-kube-api-access-rfgft" (OuterVolumeSpecName: "kube-api-access-rfgft") pod "9936e5dd-19e9-4fea-89a2-f4c0f202f880" (UID: "9936e5dd-19e9-4fea-89a2-f4c0f202f880"). InnerVolumeSpecName "kube-api-access-rfgft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:36:30 crc kubenswrapper[4871]: I1007 23:36:30.828952 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfgft\" (UniqueName: \"kubernetes.io/projected/9936e5dd-19e9-4fea-89a2-f4c0f202f880-kube-api-access-rfgft\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:31 crc kubenswrapper[4871]: I1007 23:36:31.296005 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mldkt" event={"ID":"9936e5dd-19e9-4fea-89a2-f4c0f202f880","Type":"ContainerDied","Data":"5ad0558a040d2f1ae1622edbee0dd430b3234e259933a1446d1329363b2eb79c"} Oct 07 23:36:31 crc kubenswrapper[4871]: I1007 23:36:31.296064 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ad0558a040d2f1ae1622edbee0dd430b3234e259933a1446d1329363b2eb79c" Oct 07 23:36:31 crc kubenswrapper[4871]: I1007 23:36:31.296142 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mldkt" Oct 07 23:36:35 crc kubenswrapper[4871]: I1007 23:36:35.982106 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:36:35 crc kubenswrapper[4871]: E1007 23:36:35.982867 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:36:37 crc kubenswrapper[4871]: I1007 23:36:37.791621 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-001f-account-create-8vn69"] Oct 07 23:36:37 crc kubenswrapper[4871]: E1007 23:36:37.792030 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9936e5dd-19e9-4fea-89a2-f4c0f202f880" containerName="mariadb-database-create" Oct 07 23:36:37 crc kubenswrapper[4871]: I1007 23:36:37.792042 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9936e5dd-19e9-4fea-89a2-f4c0f202f880" containerName="mariadb-database-create" Oct 07 23:36:37 crc kubenswrapper[4871]: I1007 23:36:37.792240 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9936e5dd-19e9-4fea-89a2-f4c0f202f880" containerName="mariadb-database-create" Oct 07 23:36:37 crc kubenswrapper[4871]: I1007 23:36:37.792852 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-001f-account-create-8vn69" Oct 07 23:36:37 crc kubenswrapper[4871]: I1007 23:36:37.795949 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 07 23:36:37 crc kubenswrapper[4871]: I1007 23:36:37.802214 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-001f-account-create-8vn69"] Oct 07 23:36:37 crc kubenswrapper[4871]: I1007 23:36:37.856631 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkfbc\" (UniqueName: \"kubernetes.io/projected/6376877b-2033-41cb-832c-9e9740873964-kube-api-access-kkfbc\") pod \"keystone-001f-account-create-8vn69\" (UID: \"6376877b-2033-41cb-832c-9e9740873964\") " pod="openstack/keystone-001f-account-create-8vn69" Oct 07 23:36:37 crc kubenswrapper[4871]: I1007 23:36:37.958616 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkfbc\" (UniqueName: \"kubernetes.io/projected/6376877b-2033-41cb-832c-9e9740873964-kube-api-access-kkfbc\") pod \"keystone-001f-account-create-8vn69\" (UID: \"6376877b-2033-41cb-832c-9e9740873964\") " pod="openstack/keystone-001f-account-create-8vn69" Oct 07 23:36:37 crc kubenswrapper[4871]: I1007 23:36:37.984588 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkfbc\" (UniqueName: \"kubernetes.io/projected/6376877b-2033-41cb-832c-9e9740873964-kube-api-access-kkfbc\") pod \"keystone-001f-account-create-8vn69\" (UID: \"6376877b-2033-41cb-832c-9e9740873964\") " pod="openstack/keystone-001f-account-create-8vn69" Oct 07 23:36:38 crc kubenswrapper[4871]: I1007 23:36:38.082605 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 07 23:36:38 crc kubenswrapper[4871]: I1007 23:36:38.126756 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-001f-account-create-8vn69" Oct 07 23:36:38 crc kubenswrapper[4871]: I1007 23:36:38.655614 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-001f-account-create-8vn69"] Oct 07 23:36:38 crc kubenswrapper[4871]: W1007 23:36:38.658568 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6376877b_2033_41cb_832c_9e9740873964.slice/crio-449c102d38b1dca9fc4a1166ea71d28c413dba6faa6c6618481f7ef82506379d WatchSource:0}: Error finding container 449c102d38b1dca9fc4a1166ea71d28c413dba6faa6c6618481f7ef82506379d: Status 404 returned error can't find the container with id 449c102d38b1dca9fc4a1166ea71d28c413dba6faa6c6618481f7ef82506379d Oct 07 23:36:39 crc kubenswrapper[4871]: I1007 23:36:39.386128 4871 generic.go:334] "Generic (PLEG): container finished" podID="6376877b-2033-41cb-832c-9e9740873964" containerID="90f221112885821cd8b74c5c3e4ff10f80c3c23b867c710289404d1c9753e5e4" exitCode=0 Oct 07 23:36:39 crc kubenswrapper[4871]: I1007 23:36:39.386258 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-001f-account-create-8vn69" event={"ID":"6376877b-2033-41cb-832c-9e9740873964","Type":"ContainerDied","Data":"90f221112885821cd8b74c5c3e4ff10f80c3c23b867c710289404d1c9753e5e4"} Oct 07 23:36:39 crc kubenswrapper[4871]: I1007 23:36:39.386649 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-001f-account-create-8vn69" event={"ID":"6376877b-2033-41cb-832c-9e9740873964","Type":"ContainerStarted","Data":"449c102d38b1dca9fc4a1166ea71d28c413dba6faa6c6618481f7ef82506379d"} Oct 07 23:36:40 crc kubenswrapper[4871]: I1007 23:36:40.779936 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-001f-account-create-8vn69" Oct 07 23:36:40 crc kubenswrapper[4871]: I1007 23:36:40.918105 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkfbc\" (UniqueName: \"kubernetes.io/projected/6376877b-2033-41cb-832c-9e9740873964-kube-api-access-kkfbc\") pod \"6376877b-2033-41cb-832c-9e9740873964\" (UID: \"6376877b-2033-41cb-832c-9e9740873964\") " Oct 07 23:36:40 crc kubenswrapper[4871]: I1007 23:36:40.928560 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6376877b-2033-41cb-832c-9e9740873964-kube-api-access-kkfbc" (OuterVolumeSpecName: "kube-api-access-kkfbc") pod "6376877b-2033-41cb-832c-9e9740873964" (UID: "6376877b-2033-41cb-832c-9e9740873964"). InnerVolumeSpecName "kube-api-access-kkfbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:36:41 crc kubenswrapper[4871]: I1007 23:36:41.020368 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkfbc\" (UniqueName: \"kubernetes.io/projected/6376877b-2033-41cb-832c-9e9740873964-kube-api-access-kkfbc\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:41 crc kubenswrapper[4871]: I1007 23:36:41.410975 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-001f-account-create-8vn69" event={"ID":"6376877b-2033-41cb-832c-9e9740873964","Type":"ContainerDied","Data":"449c102d38b1dca9fc4a1166ea71d28c413dba6faa6c6618481f7ef82506379d"} Oct 07 23:36:41 crc kubenswrapper[4871]: I1007 23:36:41.411028 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="449c102d38b1dca9fc4a1166ea71d28c413dba6faa6c6618481f7ef82506379d" Oct 07 23:36:41 crc kubenswrapper[4871]: I1007 23:36:41.411595 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-001f-account-create-8vn69" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.171309 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-f546f"] Oct 07 23:36:43 crc kubenswrapper[4871]: E1007 23:36:43.172863 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6376877b-2033-41cb-832c-9e9740873964" containerName="mariadb-account-create" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.172886 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6376877b-2033-41cb-832c-9e9740873964" containerName="mariadb-account-create" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.173114 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="6376877b-2033-41cb-832c-9e9740873964" containerName="mariadb-account-create" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.174123 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.176509 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gr7cv" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.176539 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.176767 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.176780 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.182831 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-f546f"] Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.360230 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwtwt\" (UniqueName: \"kubernetes.io/projected/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-kube-api-access-xwtwt\") pod \"keystone-db-sync-f546f\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.360345 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-combined-ca-bundle\") pod \"keystone-db-sync-f546f\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.360383 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-config-data\") pod \"keystone-db-sync-f546f\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.462436 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwtwt\" (UniqueName: \"kubernetes.io/projected/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-kube-api-access-xwtwt\") pod \"keystone-db-sync-f546f\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.462599 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-combined-ca-bundle\") pod \"keystone-db-sync-f546f\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.462652 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-config-data\") pod \"keystone-db-sync-f546f\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.472713 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-config-data\") pod \"keystone-db-sync-f546f\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.477113 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-combined-ca-bundle\") pod \"keystone-db-sync-f546f\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.482524 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwtwt\" (UniqueName: \"kubernetes.io/projected/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-kube-api-access-xwtwt\") pod \"keystone-db-sync-f546f\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:43 crc kubenswrapper[4871]: I1007 23:36:43.498619 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:44 crc kubenswrapper[4871]: I1007 23:36:44.002722 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-f546f"] Oct 07 23:36:44 crc kubenswrapper[4871]: I1007 23:36:44.441269 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-f546f" event={"ID":"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48","Type":"ContainerStarted","Data":"c21ce527dcbbdccc9c90446fb3706cecff68b7e4d689683367a00c6cbb1ed7a8"} Oct 07 23:36:44 crc kubenswrapper[4871]: I1007 23:36:44.441617 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-f546f" event={"ID":"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48","Type":"ContainerStarted","Data":"20d2d29b1bff365d1849aa0223d70b9571238321195b4fb89f7dde5187beb6e1"} Oct 07 23:36:44 crc kubenswrapper[4871]: I1007 23:36:44.466328 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-f546f" podStartSLOduration=1.466294314 podStartE2EDuration="1.466294314s" podCreationTimestamp="2025-10-07 23:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:44.46119403 +0000 UTC m=+5278.263892143" watchObservedRunningTime="2025-10-07 23:36:44.466294314 +0000 UTC m=+5278.268992387" Oct 07 23:36:46 crc kubenswrapper[4871]: I1007 23:36:46.467225 4871 generic.go:334] "Generic (PLEG): container finished" podID="4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48" containerID="c21ce527dcbbdccc9c90446fb3706cecff68b7e4d689683367a00c6cbb1ed7a8" exitCode=0 Oct 07 23:36:46 crc kubenswrapper[4871]: I1007 23:36:46.467334 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-f546f" event={"ID":"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48","Type":"ContainerDied","Data":"c21ce527dcbbdccc9c90446fb3706cecff68b7e4d689683367a00c6cbb1ed7a8"} Oct 07 23:36:47 crc kubenswrapper[4871]: I1007 23:36:47.880003 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.052271 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwtwt\" (UniqueName: \"kubernetes.io/projected/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-kube-api-access-xwtwt\") pod \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.052996 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-combined-ca-bundle\") pod \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.053175 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-config-data\") pod \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\" (UID: \"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48\") " Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.064905 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-kube-api-access-xwtwt" (OuterVolumeSpecName: "kube-api-access-xwtwt") pod "4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48" (UID: "4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48"). InnerVolumeSpecName "kube-api-access-xwtwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.093576 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48" (UID: "4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.120886 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-config-data" (OuterVolumeSpecName: "config-data") pod "4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48" (UID: "4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.155621 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.155675 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.155724 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwtwt\" (UniqueName: \"kubernetes.io/projected/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48-kube-api-access-xwtwt\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.491462 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-f546f" event={"ID":"4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48","Type":"ContainerDied","Data":"20d2d29b1bff365d1849aa0223d70b9571238321195b4fb89f7dde5187beb6e1"} Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.491862 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20d2d29b1bff365d1849aa0223d70b9571238321195b4fb89f7dde5187beb6e1" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.491544 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-f546f" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.709809 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f5d545f5f-gps7v"] Oct 07 23:36:48 crc kubenswrapper[4871]: E1007 23:36:48.710477 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48" containerName="keystone-db-sync" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.710490 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48" containerName="keystone-db-sync" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.710784 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48" containerName="keystone-db-sync" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.712085 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.745353 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f5d545f5f-gps7v"] Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.754022 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ww65n"] Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.755387 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.758606 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.758677 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.758851 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.759143 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gr7cv" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.759775 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ww65n"] Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.869480 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-credential-keys\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.869553 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-combined-ca-bundle\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.869586 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-scripts\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.869680 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4fsz\" (UniqueName: \"kubernetes.io/projected/98667f31-9a8f-478f-9be8-227e60b93ced-kube-api-access-n4fsz\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.869722 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-dns-svc\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.869775 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-nb\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.869868 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-sb\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.869900 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf8pc\" (UniqueName: \"kubernetes.io/projected/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-kube-api-access-bf8pc\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.869963 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-config\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.870024 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-config-data\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.870083 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-fernet-keys\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971234 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-combined-ca-bundle\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971289 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-scripts\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971322 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4fsz\" (UniqueName: \"kubernetes.io/projected/98667f31-9a8f-478f-9be8-227e60b93ced-kube-api-access-n4fsz\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971344 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-dns-svc\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971380 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-nb\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971414 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-sb\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971440 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf8pc\" (UniqueName: \"kubernetes.io/projected/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-kube-api-access-bf8pc\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971475 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-config\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971530 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-config-data\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971573 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-fernet-keys\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.971610 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-credential-keys\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.972823 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-nb\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.972824 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-sb\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.973273 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-config\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.973340 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-dns-svc\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.976412 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-credential-keys\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.976694 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-combined-ca-bundle\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.977659 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-fernet-keys\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.978730 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-config-data\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.979119 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-scripts\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.982661 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:36:48 crc kubenswrapper[4871]: E1007 23:36:48.982945 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.990409 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4fsz\" (UniqueName: \"kubernetes.io/projected/98667f31-9a8f-478f-9be8-227e60b93ced-kube-api-access-n4fsz\") pod \"dnsmasq-dns-7f5d545f5f-gps7v\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:48 crc kubenswrapper[4871]: I1007 23:36:48.994609 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf8pc\" (UniqueName: \"kubernetes.io/projected/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-kube-api-access-bf8pc\") pod \"keystone-bootstrap-ww65n\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:49 crc kubenswrapper[4871]: I1007 23:36:49.050335 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:49 crc kubenswrapper[4871]: I1007 23:36:49.077244 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:49 crc kubenswrapper[4871]: I1007 23:36:49.565735 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ww65n"] Oct 07 23:36:49 crc kubenswrapper[4871]: I1007 23:36:49.702969 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f5d545f5f-gps7v"] Oct 07 23:36:49 crc kubenswrapper[4871]: W1007 23:36:49.706853 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98667f31_9a8f_478f_9be8_227e60b93ced.slice/crio-43759222c9f8133b25544bc5de41d772bff808df637960d721068dc1e5d49f9d WatchSource:0}: Error finding container 43759222c9f8133b25544bc5de41d772bff808df637960d721068dc1e5d49f9d: Status 404 returned error can't find the container with id 43759222c9f8133b25544bc5de41d772bff808df637960d721068dc1e5d49f9d Oct 07 23:36:50 crc kubenswrapper[4871]: I1007 23:36:50.515004 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ww65n" event={"ID":"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c","Type":"ContainerStarted","Data":"81f06d6978c0d263f069798afa33d7e83748cc970921e9e76abe61a4d3b7e061"} Oct 07 23:36:50 crc kubenswrapper[4871]: I1007 23:36:50.515411 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ww65n" event={"ID":"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c","Type":"ContainerStarted","Data":"4384411affcf1406daee8e6d9a4c01c39dd5c8d85d34806c2d4caa006ed0f4ec"} Oct 07 23:36:50 crc kubenswrapper[4871]: I1007 23:36:50.518149 4871 generic.go:334] "Generic (PLEG): container finished" podID="98667f31-9a8f-478f-9be8-227e60b93ced" containerID="3253bf3df816aa3542e3c3d3afed8e44a89919ab2e13c4cf8ba2ed863e0cb2f7" exitCode=0 Oct 07 23:36:50 crc kubenswrapper[4871]: I1007 23:36:50.518241 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" event={"ID":"98667f31-9a8f-478f-9be8-227e60b93ced","Type":"ContainerDied","Data":"3253bf3df816aa3542e3c3d3afed8e44a89919ab2e13c4cf8ba2ed863e0cb2f7"} Oct 07 23:36:50 crc kubenswrapper[4871]: I1007 23:36:50.518273 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" event={"ID":"98667f31-9a8f-478f-9be8-227e60b93ced","Type":"ContainerStarted","Data":"43759222c9f8133b25544bc5de41d772bff808df637960d721068dc1e5d49f9d"} Oct 07 23:36:50 crc kubenswrapper[4871]: I1007 23:36:50.587019 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ww65n" podStartSLOduration=2.586990046 podStartE2EDuration="2.586990046s" podCreationTimestamp="2025-10-07 23:36:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:50.54433798 +0000 UTC m=+5284.347036083" watchObservedRunningTime="2025-10-07 23:36:50.586990046 +0000 UTC m=+5284.389688159" Oct 07 23:36:51 crc kubenswrapper[4871]: I1007 23:36:51.533761 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" event={"ID":"98667f31-9a8f-478f-9be8-227e60b93ced","Type":"ContainerStarted","Data":"e63ab012562588b6176ccfa30e05ce55083313fd4834e5b265207c858bcbe2f4"} Oct 07 23:36:51 crc kubenswrapper[4871]: I1007 23:36:51.534232 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:51 crc kubenswrapper[4871]: I1007 23:36:51.572195 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" podStartSLOduration=3.572153454 podStartE2EDuration="3.572153454s" podCreationTimestamp="2025-10-07 23:36:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:51.561843592 +0000 UTC m=+5285.364541695" watchObservedRunningTime="2025-10-07 23:36:51.572153454 +0000 UTC m=+5285.374851527" Oct 07 23:36:53 crc kubenswrapper[4871]: I1007 23:36:53.552981 4871 generic.go:334] "Generic (PLEG): container finished" podID="5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" containerID="81f06d6978c0d263f069798afa33d7e83748cc970921e9e76abe61a4d3b7e061" exitCode=0 Oct 07 23:36:53 crc kubenswrapper[4871]: I1007 23:36:53.553077 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ww65n" event={"ID":"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c","Type":"ContainerDied","Data":"81f06d6978c0d263f069798afa33d7e83748cc970921e9e76abe61a4d3b7e061"} Oct 07 23:36:54 crc kubenswrapper[4871]: I1007 23:36:54.999499 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.190829 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-fernet-keys\") pod \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.190979 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf8pc\" (UniqueName: \"kubernetes.io/projected/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-kube-api-access-bf8pc\") pod \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.194270 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-config-data\") pod \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.195861 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-scripts\") pod \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.195961 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-combined-ca-bundle\") pod \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.196138 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-credential-keys\") pod \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\" (UID: \"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c\") " Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.199205 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-scripts" (OuterVolumeSpecName: "scripts") pod "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" (UID: "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.199924 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" (UID: "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.199975 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" (UID: "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.201216 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-kube-api-access-bf8pc" (OuterVolumeSpecName: "kube-api-access-bf8pc") pod "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" (UID: "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c"). InnerVolumeSpecName "kube-api-access-bf8pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.239266 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-config-data" (OuterVolumeSpecName: "config-data") pod "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" (UID: "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.242989 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" (UID: "5a3b27c8-0988-4d78-bed5-6d6e504f0c4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.300570 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.300635 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.300665 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.300696 4871 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.300724 4871 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.300748 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf8pc\" (UniqueName: \"kubernetes.io/projected/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c-kube-api-access-bf8pc\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.580452 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ww65n" event={"ID":"5a3b27c8-0988-4d78-bed5-6d6e504f0c4c","Type":"ContainerDied","Data":"4384411affcf1406daee8e6d9a4c01c39dd5c8d85d34806c2d4caa006ed0f4ec"} Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.580529 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4384411affcf1406daee8e6d9a4c01c39dd5c8d85d34806c2d4caa006ed0f4ec" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.580531 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ww65n" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.676611 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ww65n"] Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.688251 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ww65n"] Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.761514 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-tm4p2"] Oct 07 23:36:55 crc kubenswrapper[4871]: E1007 23:36:55.762098 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" containerName="keystone-bootstrap" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.762132 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" containerName="keystone-bootstrap" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.762451 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" containerName="keystone-bootstrap" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.763400 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.770499 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tm4p2"] Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.772469 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.772854 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.773042 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gr7cv" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.773307 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.910032 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf55l\" (UniqueName: \"kubernetes.io/projected/84322a3a-2c29-49e7-a689-4ba61d8916c1-kube-api-access-wf55l\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.910446 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-credential-keys\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.910720 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-combined-ca-bundle\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.910787 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-fernet-keys\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.910974 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-scripts\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:55 crc kubenswrapper[4871]: I1007 23:36:55.911010 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-config-data\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.013610 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-combined-ca-bundle\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.013740 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-fernet-keys\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.013951 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-scripts\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.014728 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-config-data\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.014843 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf55l\" (UniqueName: \"kubernetes.io/projected/84322a3a-2c29-49e7-a689-4ba61d8916c1-kube-api-access-wf55l\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.014998 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-credential-keys\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.018633 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-combined-ca-bundle\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.019278 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-fernet-keys\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.019414 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-scripts\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.028162 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-credential-keys\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.028856 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-config-data\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.036900 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf55l\" (UniqueName: \"kubernetes.io/projected/84322a3a-2c29-49e7-a689-4ba61d8916c1-kube-api-access-wf55l\") pod \"keystone-bootstrap-tm4p2\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.093634 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.462482 4871 scope.go:117] "RemoveContainer" containerID="395c971e87ea1dcd6f5b65a51c13b1b9fad8615c039fe09d30696e4461bf8625" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.494436 4871 scope.go:117] "RemoveContainer" containerID="83725a48ff96b0eb167064f5d201a58371e812a6dd6fa3dab6e28e2a8a29b846" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.559224 4871 scope.go:117] "RemoveContainer" containerID="0ed394a9bfeb0ddda0005ec76d4e535d2cdb22dce1d00920901b1b01fa6ef109" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.597394 4871 scope.go:117] "RemoveContainer" containerID="dbac19c9f543e76ee7d9dece2daaf599ffdd5963c4dd867ef1e9421c428b7000" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.606774 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tm4p2"] Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.627125 4871 scope.go:117] "RemoveContainer" containerID="a857283ba9648d68b58096f8ff8568fa526f8a8ddd8d2fd710436012ceaaef32" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.661212 4871 scope.go:117] "RemoveContainer" containerID="8718a9f20f2928e2408853b41fe9d6cd12527ccfb70e0b57936a5b6f6cfa37a8" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.697779 4871 scope.go:117] "RemoveContainer" containerID="0be9dc2bc1f5930844773b5fefb20ed7ef9609d1b131bba464d11f5e77a13c0b" Oct 07 23:36:56 crc kubenswrapper[4871]: I1007 23:36:56.996240 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a3b27c8-0988-4d78-bed5-6d6e504f0c4c" path="/var/lib/kubelet/pods/5a3b27c8-0988-4d78-bed5-6d6e504f0c4c/volumes" Oct 07 23:36:57 crc kubenswrapper[4871]: I1007 23:36:57.616355 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tm4p2" event={"ID":"84322a3a-2c29-49e7-a689-4ba61d8916c1","Type":"ContainerStarted","Data":"3b157e4a5da5c53c907cc20fdfbbab34137d9981a2ebbf608f65ba73fc44b28b"} Oct 07 23:36:57 crc kubenswrapper[4871]: I1007 23:36:57.616920 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tm4p2" event={"ID":"84322a3a-2c29-49e7-a689-4ba61d8916c1","Type":"ContainerStarted","Data":"385e98d45790c53ba0002fbe3c28c35b8edae59f632f67ec98356b6c48f45c58"} Oct 07 23:36:57 crc kubenswrapper[4871]: I1007 23:36:57.647329 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-tm4p2" podStartSLOduration=2.647308533 podStartE2EDuration="2.647308533s" podCreationTimestamp="2025-10-07 23:36:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:36:57.640344919 +0000 UTC m=+5291.443043042" watchObservedRunningTime="2025-10-07 23:36:57.647308533 +0000 UTC m=+5291.450006616" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.052107 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.134313 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69765ddff5-59x29"] Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.134868 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69765ddff5-59x29" podUID="501ccea2-62f3-426b-91cd-3d1423944ef0" containerName="dnsmasq-dns" containerID="cri-o://d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d" gracePeriod=10 Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.589778 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.636213 4871 generic.go:334] "Generic (PLEG): container finished" podID="84322a3a-2c29-49e7-a689-4ba61d8916c1" containerID="3b157e4a5da5c53c907cc20fdfbbab34137d9981a2ebbf608f65ba73fc44b28b" exitCode=0 Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.636279 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tm4p2" event={"ID":"84322a3a-2c29-49e7-a689-4ba61d8916c1","Type":"ContainerDied","Data":"3b157e4a5da5c53c907cc20fdfbbab34137d9981a2ebbf608f65ba73fc44b28b"} Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.638774 4871 generic.go:334] "Generic (PLEG): container finished" podID="501ccea2-62f3-426b-91cd-3d1423944ef0" containerID="d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d" exitCode=0 Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.638829 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69765ddff5-59x29" event={"ID":"501ccea2-62f3-426b-91cd-3d1423944ef0","Type":"ContainerDied","Data":"d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d"} Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.638851 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69765ddff5-59x29" event={"ID":"501ccea2-62f3-426b-91cd-3d1423944ef0","Type":"ContainerDied","Data":"7a66befe648211bf82eeb55fefb4d782fe30bbec48c1f13aefd909964c196da3"} Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.638872 4871 scope.go:117] "RemoveContainer" containerID="d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.638980 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69765ddff5-59x29" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.680118 4871 scope.go:117] "RemoveContainer" containerID="4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.700737 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-nb\") pod \"501ccea2-62f3-426b-91cd-3d1423944ef0\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.701277 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-dns-svc\") pod \"501ccea2-62f3-426b-91cd-3d1423944ef0\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.701340 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7k24\" (UniqueName: \"kubernetes.io/projected/501ccea2-62f3-426b-91cd-3d1423944ef0-kube-api-access-r7k24\") pod \"501ccea2-62f3-426b-91cd-3d1423944ef0\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.701378 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-config\") pod \"501ccea2-62f3-426b-91cd-3d1423944ef0\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.701532 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-sb\") pod \"501ccea2-62f3-426b-91cd-3d1423944ef0\" (UID: \"501ccea2-62f3-426b-91cd-3d1423944ef0\") " Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.708122 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501ccea2-62f3-426b-91cd-3d1423944ef0-kube-api-access-r7k24" (OuterVolumeSpecName: "kube-api-access-r7k24") pod "501ccea2-62f3-426b-91cd-3d1423944ef0" (UID: "501ccea2-62f3-426b-91cd-3d1423944ef0"). InnerVolumeSpecName "kube-api-access-r7k24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.710239 4871 scope.go:117] "RemoveContainer" containerID="d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d" Oct 07 23:36:59 crc kubenswrapper[4871]: E1007 23:36:59.710632 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d\": container with ID starting with d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d not found: ID does not exist" containerID="d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.710665 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d"} err="failed to get container status \"d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d\": rpc error: code = NotFound desc = could not find container \"d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d\": container with ID starting with d5b3e9fb31129672daba68e964fb88cac0ccc9767733f25fbc1b0d3765dba68d not found: ID does not exist" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.710690 4871 scope.go:117] "RemoveContainer" containerID="4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e" Oct 07 23:36:59 crc kubenswrapper[4871]: E1007 23:36:59.710953 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e\": container with ID starting with 4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e not found: ID does not exist" containerID="4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.710981 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e"} err="failed to get container status \"4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e\": rpc error: code = NotFound desc = could not find container \"4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e\": container with ID starting with 4ca8bb06bf1c63f4bc93ed2488c07308d907a734cbd21f54a4e6faea4576251e not found: ID does not exist" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.737981 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-config" (OuterVolumeSpecName: "config") pod "501ccea2-62f3-426b-91cd-3d1423944ef0" (UID: "501ccea2-62f3-426b-91cd-3d1423944ef0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.741142 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "501ccea2-62f3-426b-91cd-3d1423944ef0" (UID: "501ccea2-62f3-426b-91cd-3d1423944ef0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.746954 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "501ccea2-62f3-426b-91cd-3d1423944ef0" (UID: "501ccea2-62f3-426b-91cd-3d1423944ef0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.751054 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "501ccea2-62f3-426b-91cd-3d1423944ef0" (UID: "501ccea2-62f3-426b-91cd-3d1423944ef0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.802436 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.802461 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.802472 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7k24\" (UniqueName: \"kubernetes.io/projected/501ccea2-62f3-426b-91cd-3d1423944ef0-kube-api-access-r7k24\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.802484 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.802492 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501ccea2-62f3-426b-91cd-3d1423944ef0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.979577 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69765ddff5-59x29"] Oct 07 23:36:59 crc kubenswrapper[4871]: I1007 23:36:59.987467 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69765ddff5-59x29"] Oct 07 23:37:00 crc kubenswrapper[4871]: I1007 23:37:00.995007 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501ccea2-62f3-426b-91cd-3d1423944ef0" path="/var/lib/kubelet/pods/501ccea2-62f3-426b-91cd-3d1423944ef0/volumes" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.011447 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.026033 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-credential-keys\") pod \"84322a3a-2c29-49e7-a689-4ba61d8916c1\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.026131 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-fernet-keys\") pod \"84322a3a-2c29-49e7-a689-4ba61d8916c1\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.026201 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-combined-ca-bundle\") pod \"84322a3a-2c29-49e7-a689-4ba61d8916c1\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.026295 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-scripts\") pod \"84322a3a-2c29-49e7-a689-4ba61d8916c1\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.026357 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-config-data\") pod \"84322a3a-2c29-49e7-a689-4ba61d8916c1\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.026418 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf55l\" (UniqueName: \"kubernetes.io/projected/84322a3a-2c29-49e7-a689-4ba61d8916c1-kube-api-access-wf55l\") pod \"84322a3a-2c29-49e7-a689-4ba61d8916c1\" (UID: \"84322a3a-2c29-49e7-a689-4ba61d8916c1\") " Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.036391 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-scripts" (OuterVolumeSpecName: "scripts") pod "84322a3a-2c29-49e7-a689-4ba61d8916c1" (UID: "84322a3a-2c29-49e7-a689-4ba61d8916c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.036441 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "84322a3a-2c29-49e7-a689-4ba61d8916c1" (UID: "84322a3a-2c29-49e7-a689-4ba61d8916c1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.036616 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84322a3a-2c29-49e7-a689-4ba61d8916c1-kube-api-access-wf55l" (OuterVolumeSpecName: "kube-api-access-wf55l") pod "84322a3a-2c29-49e7-a689-4ba61d8916c1" (UID: "84322a3a-2c29-49e7-a689-4ba61d8916c1"). InnerVolumeSpecName "kube-api-access-wf55l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.037175 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "84322a3a-2c29-49e7-a689-4ba61d8916c1" (UID: "84322a3a-2c29-49e7-a689-4ba61d8916c1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.056232 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84322a3a-2c29-49e7-a689-4ba61d8916c1" (UID: "84322a3a-2c29-49e7-a689-4ba61d8916c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.058014 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-config-data" (OuterVolumeSpecName: "config-data") pod "84322a3a-2c29-49e7-a689-4ba61d8916c1" (UID: "84322a3a-2c29-49e7-a689-4ba61d8916c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.128387 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf55l\" (UniqueName: \"kubernetes.io/projected/84322a3a-2c29-49e7-a689-4ba61d8916c1-kube-api-access-wf55l\") on node \"crc\" DevicePath \"\"" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.128437 4871 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.128454 4871 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.128470 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.128484 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.128496 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84322a3a-2c29-49e7-a689-4ba61d8916c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.666610 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tm4p2" event={"ID":"84322a3a-2c29-49e7-a689-4ba61d8916c1","Type":"ContainerDied","Data":"385e98d45790c53ba0002fbe3c28c35b8edae59f632f67ec98356b6c48f45c58"} Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.666677 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="385e98d45790c53ba0002fbe3c28c35b8edae59f632f67ec98356b6c48f45c58" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.667410 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tm4p2" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.764413 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-dbc89c896-7kvkt"] Oct 07 23:37:01 crc kubenswrapper[4871]: E1007 23:37:01.765133 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501ccea2-62f3-426b-91cd-3d1423944ef0" containerName="init" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.765166 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="501ccea2-62f3-426b-91cd-3d1423944ef0" containerName="init" Oct 07 23:37:01 crc kubenswrapper[4871]: E1007 23:37:01.765195 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84322a3a-2c29-49e7-a689-4ba61d8916c1" containerName="keystone-bootstrap" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.765207 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="84322a3a-2c29-49e7-a689-4ba61d8916c1" containerName="keystone-bootstrap" Oct 07 23:37:01 crc kubenswrapper[4871]: E1007 23:37:01.765256 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501ccea2-62f3-426b-91cd-3d1423944ef0" containerName="dnsmasq-dns" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.765269 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="501ccea2-62f3-426b-91cd-3d1423944ef0" containerName="dnsmasq-dns" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.765529 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="501ccea2-62f3-426b-91cd-3d1423944ef0" containerName="dnsmasq-dns" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.765560 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="84322a3a-2c29-49e7-a689-4ba61d8916c1" containerName="keystone-bootstrap" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.766667 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.769729 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gr7cv" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.770103 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.770892 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.770910 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.793402 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-dbc89c896-7kvkt"] Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.840421 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6srlv\" (UniqueName: \"kubernetes.io/projected/3500a587-fccc-460c-9848-520a06582186-kube-api-access-6srlv\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.840505 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-config-data\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.840536 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-scripts\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.840551 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-combined-ca-bundle\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.840572 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-credential-keys\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.840838 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-fernet-keys\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.942986 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-config-data\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.943126 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-scripts\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.943174 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-combined-ca-bundle\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.943218 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-credential-keys\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.943303 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-fernet-keys\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.943395 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6srlv\" (UniqueName: \"kubernetes.io/projected/3500a587-fccc-460c-9848-520a06582186-kube-api-access-6srlv\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.947431 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-credential-keys\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.948287 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-combined-ca-bundle\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.948863 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-config-data\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.949411 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-scripts\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.952402 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3500a587-fccc-460c-9848-520a06582186-fernet-keys\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:01 crc kubenswrapper[4871]: I1007 23:37:01.963622 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6srlv\" (UniqueName: \"kubernetes.io/projected/3500a587-fccc-460c-9848-520a06582186-kube-api-access-6srlv\") pod \"keystone-dbc89c896-7kvkt\" (UID: \"3500a587-fccc-460c-9848-520a06582186\") " pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:02 crc kubenswrapper[4871]: I1007 23:37:02.089509 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:02 crc kubenswrapper[4871]: I1007 23:37:02.563564 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-dbc89c896-7kvkt"] Oct 07 23:37:02 crc kubenswrapper[4871]: I1007 23:37:02.676755 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-dbc89c896-7kvkt" event={"ID":"3500a587-fccc-460c-9848-520a06582186","Type":"ContainerStarted","Data":"c725a8dc4082f804376c5b024a2d18aa72609ea13e5bf8532d14ff94f466deb3"} Oct 07 23:37:03 crc kubenswrapper[4871]: I1007 23:37:03.685225 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-dbc89c896-7kvkt" event={"ID":"3500a587-fccc-460c-9848-520a06582186","Type":"ContainerStarted","Data":"97d6c12f30f97accf093bb807a62a56b117a0af35a7154c1a0b56624897d8cd8"} Oct 07 23:37:03 crc kubenswrapper[4871]: I1007 23:37:03.685611 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:03 crc kubenswrapper[4871]: I1007 23:37:03.713725 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-dbc89c896-7kvkt" podStartSLOduration=2.713700652 podStartE2EDuration="2.713700652s" podCreationTimestamp="2025-10-07 23:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:37:03.702388603 +0000 UTC m=+5297.505086736" watchObservedRunningTime="2025-10-07 23:37:03.713700652 +0000 UTC m=+5297.516398745" Oct 07 23:37:03 crc kubenswrapper[4871]: I1007 23:37:03.982695 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:37:03 crc kubenswrapper[4871]: E1007 23:37:03.983256 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.577344 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zp4nr"] Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.580426 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.584557 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zp4nr"] Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.778600 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-catalog-content\") pod \"community-operators-zp4nr\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.779562 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8b6h\" (UniqueName: \"kubernetes.io/projected/6a698c82-034c-45b0-b1df-bb25617431e5-kube-api-access-q8b6h\") pod \"community-operators-zp4nr\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.779679 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-utilities\") pod \"community-operators-zp4nr\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.881354 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-catalog-content\") pod \"community-operators-zp4nr\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.881476 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8b6h\" (UniqueName: \"kubernetes.io/projected/6a698c82-034c-45b0-b1df-bb25617431e5-kube-api-access-q8b6h\") pod \"community-operators-zp4nr\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.881526 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-utilities\") pod \"community-operators-zp4nr\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.882021 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-catalog-content\") pod \"community-operators-zp4nr\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.882055 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-utilities\") pod \"community-operators-zp4nr\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:13 crc kubenswrapper[4871]: I1007 23:37:13.916001 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8b6h\" (UniqueName: \"kubernetes.io/projected/6a698c82-034c-45b0-b1df-bb25617431e5-kube-api-access-q8b6h\") pod \"community-operators-zp4nr\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:14 crc kubenswrapper[4871]: I1007 23:37:14.203259 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:14 crc kubenswrapper[4871]: I1007 23:37:14.472822 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zp4nr"] Oct 07 23:37:14 crc kubenswrapper[4871]: I1007 23:37:14.811200 4871 generic.go:334] "Generic (PLEG): container finished" podID="6a698c82-034c-45b0-b1df-bb25617431e5" containerID="a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19" exitCode=0 Oct 07 23:37:14 crc kubenswrapper[4871]: I1007 23:37:14.811291 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp4nr" event={"ID":"6a698c82-034c-45b0-b1df-bb25617431e5","Type":"ContainerDied","Data":"a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19"} Oct 07 23:37:14 crc kubenswrapper[4871]: I1007 23:37:14.812564 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp4nr" event={"ID":"6a698c82-034c-45b0-b1df-bb25617431e5","Type":"ContainerStarted","Data":"172b1c46ff67e30b6243e8d41095016563e4f599d09e2c0301ac584a17f11896"} Oct 07 23:37:15 crc kubenswrapper[4871]: I1007 23:37:15.823229 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp4nr" event={"ID":"6a698c82-034c-45b0-b1df-bb25617431e5","Type":"ContainerStarted","Data":"676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad"} Oct 07 23:37:16 crc kubenswrapper[4871]: I1007 23:37:16.837451 4871 generic.go:334] "Generic (PLEG): container finished" podID="6a698c82-034c-45b0-b1df-bb25617431e5" containerID="676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad" exitCode=0 Oct 07 23:37:16 crc kubenswrapper[4871]: I1007 23:37:16.837524 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp4nr" event={"ID":"6a698c82-034c-45b0-b1df-bb25617431e5","Type":"ContainerDied","Data":"676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad"} Oct 07 23:37:17 crc kubenswrapper[4871]: I1007 23:37:17.852539 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp4nr" event={"ID":"6a698c82-034c-45b0-b1df-bb25617431e5","Type":"ContainerStarted","Data":"ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58"} Oct 07 23:37:17 crc kubenswrapper[4871]: I1007 23:37:17.886713 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zp4nr" podStartSLOduration=2.360775686 podStartE2EDuration="4.886683547s" podCreationTimestamp="2025-10-07 23:37:13 +0000 UTC" firstStartedPulling="2025-10-07 23:37:14.813843277 +0000 UTC m=+5308.616541390" lastFinishedPulling="2025-10-07 23:37:17.339751138 +0000 UTC m=+5311.142449251" observedRunningTime="2025-10-07 23:37:17.885213868 +0000 UTC m=+5311.687911961" watchObservedRunningTime="2025-10-07 23:37:17.886683547 +0000 UTC m=+5311.689381660" Oct 07 23:37:18 crc kubenswrapper[4871]: I1007 23:37:18.982888 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:37:19 crc kubenswrapper[4871]: I1007 23:37:19.882147 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"a7d7ec4f805a57b5cbdcea4bbeab976af5520b960d9f0ff723e036600ebfeb03"} Oct 07 23:37:24 crc kubenswrapper[4871]: I1007 23:37:24.204314 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:24 crc kubenswrapper[4871]: I1007 23:37:24.205150 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:24 crc kubenswrapper[4871]: I1007 23:37:24.283035 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:25 crc kubenswrapper[4871]: I1007 23:37:25.005206 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:25 crc kubenswrapper[4871]: I1007 23:37:25.076309 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zp4nr"] Oct 07 23:37:26 crc kubenswrapper[4871]: I1007 23:37:26.959057 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zp4nr" podUID="6a698c82-034c-45b0-b1df-bb25617431e5" containerName="registry-server" containerID="cri-o://ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58" gracePeriod=2 Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.446159 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.536634 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-catalog-content\") pod \"6a698c82-034c-45b0-b1df-bb25617431e5\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.536687 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-utilities\") pod \"6a698c82-034c-45b0-b1df-bb25617431e5\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.537754 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-utilities" (OuterVolumeSpecName: "utilities") pod "6a698c82-034c-45b0-b1df-bb25617431e5" (UID: "6a698c82-034c-45b0-b1df-bb25617431e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.537923 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8b6h\" (UniqueName: \"kubernetes.io/projected/6a698c82-034c-45b0-b1df-bb25617431e5-kube-api-access-q8b6h\") pod \"6a698c82-034c-45b0-b1df-bb25617431e5\" (UID: \"6a698c82-034c-45b0-b1df-bb25617431e5\") " Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.539182 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.547446 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a698c82-034c-45b0-b1df-bb25617431e5-kube-api-access-q8b6h" (OuterVolumeSpecName: "kube-api-access-q8b6h") pod "6a698c82-034c-45b0-b1df-bb25617431e5" (UID: "6a698c82-034c-45b0-b1df-bb25617431e5"). InnerVolumeSpecName "kube-api-access-q8b6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.607127 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a698c82-034c-45b0-b1df-bb25617431e5" (UID: "6a698c82-034c-45b0-b1df-bb25617431e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.641165 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8b6h\" (UniqueName: \"kubernetes.io/projected/6a698c82-034c-45b0-b1df-bb25617431e5-kube-api-access-q8b6h\") on node \"crc\" DevicePath \"\"" Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.641215 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a698c82-034c-45b0-b1df-bb25617431e5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.972968 4871 generic.go:334] "Generic (PLEG): container finished" podID="6a698c82-034c-45b0-b1df-bb25617431e5" containerID="ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58" exitCode=0 Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.973033 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp4nr" event={"ID":"6a698c82-034c-45b0-b1df-bb25617431e5","Type":"ContainerDied","Data":"ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58"} Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.973067 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zp4nr" Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.973084 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp4nr" event={"ID":"6a698c82-034c-45b0-b1df-bb25617431e5","Type":"ContainerDied","Data":"172b1c46ff67e30b6243e8d41095016563e4f599d09e2c0301ac584a17f11896"} Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.973116 4871 scope.go:117] "RemoveContainer" containerID="ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58" Oct 07 23:37:27 crc kubenswrapper[4871]: I1007 23:37:27.998682 4871 scope.go:117] "RemoveContainer" containerID="676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad" Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.024529 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zp4nr"] Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.046676 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zp4nr"] Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.051471 4871 scope.go:117] "RemoveContainer" containerID="a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19" Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.086261 4871 scope.go:117] "RemoveContainer" containerID="ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58" Oct 07 23:37:28 crc kubenswrapper[4871]: E1007 23:37:28.086574 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58\": container with ID starting with ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58 not found: ID does not exist" containerID="ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58" Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.086606 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58"} err="failed to get container status \"ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58\": rpc error: code = NotFound desc = could not find container \"ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58\": container with ID starting with ec6aee792a962ceb578672722e8511b37a2864d65b567a5b4a58375f05616c58 not found: ID does not exist" Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.086626 4871 scope.go:117] "RemoveContainer" containerID="676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad" Oct 07 23:37:28 crc kubenswrapper[4871]: E1007 23:37:28.087456 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad\": container with ID starting with 676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad not found: ID does not exist" containerID="676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad" Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.087500 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad"} err="failed to get container status \"676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad\": rpc error: code = NotFound desc = could not find container \"676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad\": container with ID starting with 676e333235bfdfc53896cc8f4e914a3050dd447e43260aec6e253b136f818aad not found: ID does not exist" Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.087533 4871 scope.go:117] "RemoveContainer" containerID="a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19" Oct 07 23:37:28 crc kubenswrapper[4871]: E1007 23:37:28.087986 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19\": container with ID starting with a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19 not found: ID does not exist" containerID="a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19" Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.088012 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19"} err="failed to get container status \"a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19\": rpc error: code = NotFound desc = could not find container \"a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19\": container with ID starting with a3009cb817ae8661df4fb3c11be03c8ca18e2a365ffaad6e8f380743a0f7bb19 not found: ID does not exist" Oct 07 23:37:28 crc kubenswrapper[4871]: I1007 23:37:28.998896 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a698c82-034c-45b0-b1df-bb25617431e5" path="/var/lib/kubelet/pods/6a698c82-034c-45b0-b1df-bb25617431e5/volumes" Oct 07 23:37:33 crc kubenswrapper[4871]: I1007 23:37:33.592256 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-dbc89c896-7kvkt" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.819319 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 07 23:37:37 crc kubenswrapper[4871]: E1007 23:37:37.819844 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a698c82-034c-45b0-b1df-bb25617431e5" containerName="extract-utilities" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.819866 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a698c82-034c-45b0-b1df-bb25617431e5" containerName="extract-utilities" Oct 07 23:37:37 crc kubenswrapper[4871]: E1007 23:37:37.819917 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a698c82-034c-45b0-b1df-bb25617431e5" containerName="extract-content" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.819930 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a698c82-034c-45b0-b1df-bb25617431e5" containerName="extract-content" Oct 07 23:37:37 crc kubenswrapper[4871]: E1007 23:37:37.819960 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a698c82-034c-45b0-b1df-bb25617431e5" containerName="registry-server" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.819972 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a698c82-034c-45b0-b1df-bb25617431e5" containerName="registry-server" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.820260 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a698c82-034c-45b0-b1df-bb25617431e5" containerName="registry-server" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.821137 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.824690 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.824735 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-wql7b" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.824737 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.843582 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.929088 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jglz\" (UniqueName: \"kubernetes.io/projected/15bced8d-616c-4f9b-b265-b0eb216718c8-kube-api-access-7jglz\") pod \"openstackclient\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " pod="openstack/openstackclient" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.929164 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config-secret\") pod \"openstackclient\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " pod="openstack/openstackclient" Oct 07 23:37:37 crc kubenswrapper[4871]: I1007 23:37:37.929293 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config\") pod \"openstackclient\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " pod="openstack/openstackclient" Oct 07 23:37:38 crc kubenswrapper[4871]: I1007 23:37:38.030883 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config\") pod \"openstackclient\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " pod="openstack/openstackclient" Oct 07 23:37:38 crc kubenswrapper[4871]: I1007 23:37:38.031102 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jglz\" (UniqueName: \"kubernetes.io/projected/15bced8d-616c-4f9b-b265-b0eb216718c8-kube-api-access-7jglz\") pod \"openstackclient\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " pod="openstack/openstackclient" Oct 07 23:37:38 crc kubenswrapper[4871]: I1007 23:37:38.031151 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config-secret\") pod \"openstackclient\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " pod="openstack/openstackclient" Oct 07 23:37:38 crc kubenswrapper[4871]: I1007 23:37:38.032398 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config\") pod \"openstackclient\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " pod="openstack/openstackclient" Oct 07 23:37:38 crc kubenswrapper[4871]: I1007 23:37:38.040873 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config-secret\") pod \"openstackclient\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " pod="openstack/openstackclient" Oct 07 23:37:38 crc kubenswrapper[4871]: I1007 23:37:38.069769 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jglz\" (UniqueName: \"kubernetes.io/projected/15bced8d-616c-4f9b-b265-b0eb216718c8-kube-api-access-7jglz\") pod \"openstackclient\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " pod="openstack/openstackclient" Oct 07 23:37:38 crc kubenswrapper[4871]: I1007 23:37:38.155561 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 23:37:38 crc kubenswrapper[4871]: I1007 23:37:38.686885 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 23:37:39 crc kubenswrapper[4871]: I1007 23:37:39.078957 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"15bced8d-616c-4f9b-b265-b0eb216718c8","Type":"ContainerStarted","Data":"3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e"} Oct 07 23:37:39 crc kubenswrapper[4871]: I1007 23:37:39.079409 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"15bced8d-616c-4f9b-b265-b0eb216718c8","Type":"ContainerStarted","Data":"9b498306d04fbc5a48a2eaf8ecd72e7dfc1d084de51220958784c5b88ade362d"} Oct 07 23:37:39 crc kubenswrapper[4871]: I1007 23:37:39.097618 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.097600018 podStartE2EDuration="2.097600018s" podCreationTimestamp="2025-10-07 23:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:37:39.095564064 +0000 UTC m=+5332.898262137" watchObservedRunningTime="2025-10-07 23:37:39.097600018 +0000 UTC m=+5332.900298091" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.441586 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8cnvf"] Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.444280 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.455849 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8cnvf"] Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.556663 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-catalog-content\") pod \"redhat-operators-8cnvf\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.557031 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qr94\" (UniqueName: \"kubernetes.io/projected/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-kube-api-access-4qr94\") pod \"redhat-operators-8cnvf\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.557213 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-utilities\") pod \"redhat-operators-8cnvf\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.658367 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-catalog-content\") pod \"redhat-operators-8cnvf\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.658498 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qr94\" (UniqueName: \"kubernetes.io/projected/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-kube-api-access-4qr94\") pod \"redhat-operators-8cnvf\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.658544 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-utilities\") pod \"redhat-operators-8cnvf\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.659227 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-utilities\") pod \"redhat-operators-8cnvf\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.659551 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-catalog-content\") pod \"redhat-operators-8cnvf\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.700661 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qr94\" (UniqueName: \"kubernetes.io/projected/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-kube-api-access-4qr94\") pod \"redhat-operators-8cnvf\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:33 crc kubenswrapper[4871]: I1007 23:38:33.771388 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:34 crc kubenswrapper[4871]: I1007 23:38:34.278568 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8cnvf"] Oct 07 23:38:34 crc kubenswrapper[4871]: W1007 23:38:34.282162 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48b9dd08_8514_4d36_b7a8_95bac9b16fbe.slice/crio-3c5c4e61c8720910c36bebef95360e55743d1ee2fba27e763d2d783e26569b36 WatchSource:0}: Error finding container 3c5c4e61c8720910c36bebef95360e55743d1ee2fba27e763d2d783e26569b36: Status 404 returned error can't find the container with id 3c5c4e61c8720910c36bebef95360e55743d1ee2fba27e763d2d783e26569b36 Oct 07 23:38:34 crc kubenswrapper[4871]: I1007 23:38:34.648874 4871 generic.go:334] "Generic (PLEG): container finished" podID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerID="b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064" exitCode=0 Oct 07 23:38:34 crc kubenswrapper[4871]: I1007 23:38:34.648931 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cnvf" event={"ID":"48b9dd08-8514-4d36-b7a8-95bac9b16fbe","Type":"ContainerDied","Data":"b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064"} Oct 07 23:38:34 crc kubenswrapper[4871]: I1007 23:38:34.649180 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cnvf" event={"ID":"48b9dd08-8514-4d36-b7a8-95bac9b16fbe","Type":"ContainerStarted","Data":"3c5c4e61c8720910c36bebef95360e55743d1ee2fba27e763d2d783e26569b36"} Oct 07 23:38:36 crc kubenswrapper[4871]: I1007 23:38:36.676334 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cnvf" event={"ID":"48b9dd08-8514-4d36-b7a8-95bac9b16fbe","Type":"ContainerStarted","Data":"97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b"} Oct 07 23:38:37 crc kubenswrapper[4871]: I1007 23:38:37.691648 4871 generic.go:334] "Generic (PLEG): container finished" podID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerID="97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b" exitCode=0 Oct 07 23:38:37 crc kubenswrapper[4871]: I1007 23:38:37.692065 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cnvf" event={"ID":"48b9dd08-8514-4d36-b7a8-95bac9b16fbe","Type":"ContainerDied","Data":"97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b"} Oct 07 23:38:38 crc kubenswrapper[4871]: I1007 23:38:38.705960 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cnvf" event={"ID":"48b9dd08-8514-4d36-b7a8-95bac9b16fbe","Type":"ContainerStarted","Data":"b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56"} Oct 07 23:38:43 crc kubenswrapper[4871]: I1007 23:38:43.772100 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:43 crc kubenswrapper[4871]: I1007 23:38:43.774144 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:44 crc kubenswrapper[4871]: I1007 23:38:44.837270 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8cnvf" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerName="registry-server" probeResult="failure" output=< Oct 07 23:38:44 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 23:38:44 crc kubenswrapper[4871]: > Oct 07 23:38:53 crc kubenswrapper[4871]: I1007 23:38:53.819251 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:53 crc kubenswrapper[4871]: I1007 23:38:53.851411 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8cnvf" podStartSLOduration=17.239060716 podStartE2EDuration="20.851386249s" podCreationTimestamp="2025-10-07 23:38:33 +0000 UTC" firstStartedPulling="2025-10-07 23:38:34.650555261 +0000 UTC m=+5388.453253334" lastFinishedPulling="2025-10-07 23:38:38.262880784 +0000 UTC m=+5392.065578867" observedRunningTime="2025-10-07 23:38:38.724679366 +0000 UTC m=+5392.527377449" watchObservedRunningTime="2025-10-07 23:38:53.851386249 +0000 UTC m=+5407.654084332" Oct 07 23:38:53 crc kubenswrapper[4871]: I1007 23:38:53.870309 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:54 crc kubenswrapper[4871]: I1007 23:38:54.071442 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8cnvf"] Oct 07 23:38:54 crc kubenswrapper[4871]: I1007 23:38:54.873743 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8cnvf" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerName="registry-server" containerID="cri-o://b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56" gracePeriod=2 Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.326330 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.393341 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qr94\" (UniqueName: \"kubernetes.io/projected/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-kube-api-access-4qr94\") pod \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.393422 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-utilities\") pod \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.393521 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-catalog-content\") pod \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\" (UID: \"48b9dd08-8514-4d36-b7a8-95bac9b16fbe\") " Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.394813 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-utilities" (OuterVolumeSpecName: "utilities") pod "48b9dd08-8514-4d36-b7a8-95bac9b16fbe" (UID: "48b9dd08-8514-4d36-b7a8-95bac9b16fbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.398935 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-kube-api-access-4qr94" (OuterVolumeSpecName: "kube-api-access-4qr94") pod "48b9dd08-8514-4d36-b7a8-95bac9b16fbe" (UID: "48b9dd08-8514-4d36-b7a8-95bac9b16fbe"). InnerVolumeSpecName "kube-api-access-4qr94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.477542 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48b9dd08-8514-4d36-b7a8-95bac9b16fbe" (UID: "48b9dd08-8514-4d36-b7a8-95bac9b16fbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.495744 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.495781 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.495793 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qr94\" (UniqueName: \"kubernetes.io/projected/48b9dd08-8514-4d36-b7a8-95bac9b16fbe-kube-api-access-4qr94\") on node \"crc\" DevicePath \"\"" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.890541 4871 generic.go:334] "Generic (PLEG): container finished" podID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerID="b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56" exitCode=0 Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.890633 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cnvf" event={"ID":"48b9dd08-8514-4d36-b7a8-95bac9b16fbe","Type":"ContainerDied","Data":"b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56"} Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.890761 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cnvf" event={"ID":"48b9dd08-8514-4d36-b7a8-95bac9b16fbe","Type":"ContainerDied","Data":"3c5c4e61c8720910c36bebef95360e55743d1ee2fba27e763d2d783e26569b36"} Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.890851 4871 scope.go:117] "RemoveContainer" containerID="b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.893834 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cnvf" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.913554 4871 scope.go:117] "RemoveContainer" containerID="97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.951202 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8cnvf"] Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.958940 4871 scope.go:117] "RemoveContainer" containerID="b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.963317 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8cnvf"] Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.980879 4871 scope.go:117] "RemoveContainer" containerID="b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56" Oct 07 23:38:55 crc kubenswrapper[4871]: E1007 23:38:55.981628 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56\": container with ID starting with b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56 not found: ID does not exist" containerID="b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.981691 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56"} err="failed to get container status \"b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56\": rpc error: code = NotFound desc = could not find container \"b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56\": container with ID starting with b887968de0b96aba759b320ebef19efdbb50667c37c218d2fb13ceede1d73f56 not found: ID does not exist" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.981732 4871 scope.go:117] "RemoveContainer" containerID="97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b" Oct 07 23:38:55 crc kubenswrapper[4871]: E1007 23:38:55.984210 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b\": container with ID starting with 97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b not found: ID does not exist" containerID="97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.984258 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b"} err="failed to get container status \"97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b\": rpc error: code = NotFound desc = could not find container \"97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b\": container with ID starting with 97fd6127413974bc7283f37655a90681b9ab397c5e0c34c32e2111978c5faa8b not found: ID does not exist" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.984284 4871 scope.go:117] "RemoveContainer" containerID="b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064" Oct 07 23:38:55 crc kubenswrapper[4871]: E1007 23:38:55.989962 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064\": container with ID starting with b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064 not found: ID does not exist" containerID="b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064" Oct 07 23:38:55 crc kubenswrapper[4871]: I1007 23:38:55.990068 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064"} err="failed to get container status \"b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064\": rpc error: code = NotFound desc = could not find container \"b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064\": container with ID starting with b7acd6e4818cdd6456fc65844c3f29fd479d94ec06f10d3ef14335116139b064 not found: ID does not exist" Oct 07 23:38:56 crc kubenswrapper[4871]: I1007 23:38:56.999475 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" path="/var/lib/kubelet/pods/48b9dd08-8514-4d36-b7a8-95bac9b16fbe/volumes" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.517690 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mfljv"] Oct 07 23:39:02 crc kubenswrapper[4871]: E1007 23:39:02.518519 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerName="extract-utilities" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.518532 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerName="extract-utilities" Oct 07 23:39:02 crc kubenswrapper[4871]: E1007 23:39:02.518556 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerName="registry-server" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.518562 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerName="registry-server" Oct 07 23:39:02 crc kubenswrapper[4871]: E1007 23:39:02.518571 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerName="extract-content" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.518577 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerName="extract-content" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.518741 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="48b9dd08-8514-4d36-b7a8-95bac9b16fbe" containerName="registry-server" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.519996 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.531134 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mfljv"] Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.556328 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-utilities\") pod \"redhat-marketplace-mfljv\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.556422 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-catalog-content\") pod \"redhat-marketplace-mfljv\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.556483 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gxhl\" (UniqueName: \"kubernetes.io/projected/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-kube-api-access-2gxhl\") pod \"redhat-marketplace-mfljv\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.658232 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-utilities\") pod \"redhat-marketplace-mfljv\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.658317 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-catalog-content\") pod \"redhat-marketplace-mfljv\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.658358 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gxhl\" (UniqueName: \"kubernetes.io/projected/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-kube-api-access-2gxhl\") pod \"redhat-marketplace-mfljv\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.658874 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-utilities\") pod \"redhat-marketplace-mfljv\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.658886 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-catalog-content\") pod \"redhat-marketplace-mfljv\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.677292 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gxhl\" (UniqueName: \"kubernetes.io/projected/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-kube-api-access-2gxhl\") pod \"redhat-marketplace-mfljv\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:02 crc kubenswrapper[4871]: I1007 23:39:02.844194 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:03 crc kubenswrapper[4871]: W1007 23:39:03.171662 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2dbeb7fc_74b2_4dbf_be4e_ce2570c56365.slice/crio-90002d20f267eaf586ba4e1fd76e79905e740caed37418b6bb136f044057c00f WatchSource:0}: Error finding container 90002d20f267eaf586ba4e1fd76e79905e740caed37418b6bb136f044057c00f: Status 404 returned error can't find the container with id 90002d20f267eaf586ba4e1fd76e79905e740caed37418b6bb136f044057c00f Oct 07 23:39:03 crc kubenswrapper[4871]: I1007 23:39:03.189274 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mfljv"] Oct 07 23:39:03 crc kubenswrapper[4871]: E1007 23:39:03.794763 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2dbeb7fc_74b2_4dbf_be4e_ce2570c56365.slice/crio-conmon-a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2dbeb7fc_74b2_4dbf_be4e_ce2570c56365.slice/crio-a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b.scope\": RecentStats: unable to find data in memory cache]" Oct 07 23:39:03 crc kubenswrapper[4871]: I1007 23:39:03.961626 4871 generic.go:334] "Generic (PLEG): container finished" podID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerID="a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b" exitCode=0 Oct 07 23:39:03 crc kubenswrapper[4871]: I1007 23:39:03.961673 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mfljv" event={"ID":"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365","Type":"ContainerDied","Data":"a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b"} Oct 07 23:39:03 crc kubenswrapper[4871]: I1007 23:39:03.961718 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mfljv" event={"ID":"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365","Type":"ContainerStarted","Data":"90002d20f267eaf586ba4e1fd76e79905e740caed37418b6bb136f044057c00f"} Oct 07 23:39:06 crc kubenswrapper[4871]: I1007 23:39:06.986947 4871 generic.go:334] "Generic (PLEG): container finished" podID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerID="d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de" exitCode=0 Oct 07 23:39:06 crc kubenswrapper[4871]: I1007 23:39:06.993015 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mfljv" event={"ID":"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365","Type":"ContainerDied","Data":"d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de"} Oct 07 23:39:08 crc kubenswrapper[4871]: I1007 23:39:08.003296 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mfljv" event={"ID":"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365","Type":"ContainerStarted","Data":"b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41"} Oct 07 23:39:08 crc kubenswrapper[4871]: I1007 23:39:08.028479 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mfljv" podStartSLOduration=2.350311622 podStartE2EDuration="6.028461713s" podCreationTimestamp="2025-10-07 23:39:02 +0000 UTC" firstStartedPulling="2025-10-07 23:39:03.964380594 +0000 UTC m=+5417.767078707" lastFinishedPulling="2025-10-07 23:39:07.642530685 +0000 UTC m=+5421.445228798" observedRunningTime="2025-10-07 23:39:08.02266484 +0000 UTC m=+5421.825362913" watchObservedRunningTime="2025-10-07 23:39:08.028461713 +0000 UTC m=+5421.831159786" Oct 07 23:39:12 crc kubenswrapper[4871]: I1007 23:39:12.845137 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:12 crc kubenswrapper[4871]: I1007 23:39:12.845703 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:12 crc kubenswrapper[4871]: I1007 23:39:12.895119 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:13 crc kubenswrapper[4871]: I1007 23:39:13.103461 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:13 crc kubenswrapper[4871]: I1007 23:39:13.146577 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mfljv"] Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.091011 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mfljv" podUID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerName="registry-server" containerID="cri-o://b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41" gracePeriod=2 Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.580252 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.691444 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-utilities\") pod \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.692036 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-catalog-content\") pod \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.692126 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gxhl\" (UniqueName: \"kubernetes.io/projected/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-kube-api-access-2gxhl\") pod \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\" (UID: \"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365\") " Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.692499 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-utilities" (OuterVolumeSpecName: "utilities") pod "2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" (UID: "2dbeb7fc-74b2-4dbf-be4e-ce2570c56365"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.697342 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.702106 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-kube-api-access-2gxhl" (OuterVolumeSpecName: "kube-api-access-2gxhl") pod "2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" (UID: "2dbeb7fc-74b2-4dbf-be4e-ce2570c56365"). InnerVolumeSpecName "kube-api-access-2gxhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.723605 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" (UID: "2dbeb7fc-74b2-4dbf-be4e-ce2570c56365"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.798778 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:15 crc kubenswrapper[4871]: I1007 23:39:15.798863 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gxhl\" (UniqueName: \"kubernetes.io/projected/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365-kube-api-access-2gxhl\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.106741 4871 generic.go:334] "Generic (PLEG): container finished" podID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerID="b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41" exitCode=0 Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.106870 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mfljv" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.106898 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mfljv" event={"ID":"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365","Type":"ContainerDied","Data":"b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41"} Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.108196 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mfljv" event={"ID":"2dbeb7fc-74b2-4dbf-be4e-ce2570c56365","Type":"ContainerDied","Data":"90002d20f267eaf586ba4e1fd76e79905e740caed37418b6bb136f044057c00f"} Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.108225 4871 scope.go:117] "RemoveContainer" containerID="b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.132166 4871 scope.go:117] "RemoveContainer" containerID="d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.152767 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mfljv"] Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.160178 4871 scope.go:117] "RemoveContainer" containerID="a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.171517 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mfljv"] Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.217112 4871 scope.go:117] "RemoveContainer" containerID="b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41" Oct 07 23:39:16 crc kubenswrapper[4871]: E1007 23:39:16.217822 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41\": container with ID starting with b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41 not found: ID does not exist" containerID="b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.217883 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41"} err="failed to get container status \"b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41\": rpc error: code = NotFound desc = could not find container \"b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41\": container with ID starting with b6dcced7d9e684b2dd495547e1a024de9888d1421fbe4980ff61b772ecc05a41 not found: ID does not exist" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.217937 4871 scope.go:117] "RemoveContainer" containerID="d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de" Oct 07 23:39:16 crc kubenswrapper[4871]: E1007 23:39:16.218315 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de\": container with ID starting with d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de not found: ID does not exist" containerID="d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.218365 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de"} err="failed to get container status \"d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de\": rpc error: code = NotFound desc = could not find container \"d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de\": container with ID starting with d9f5491446dc86c83fe74fecf79cd0e8740634ba1aa724bfa4bf15e4344573de not found: ID does not exist" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.218399 4871 scope.go:117] "RemoveContainer" containerID="a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b" Oct 07 23:39:16 crc kubenswrapper[4871]: E1007 23:39:16.218713 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b\": container with ID starting with a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b not found: ID does not exist" containerID="a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.218746 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b"} err="failed to get container status \"a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b\": rpc error: code = NotFound desc = could not find container \"a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b\": container with ID starting with a786484886789d9a787fb08653121edef5595e740b987d4caa4f19b85760aa4b not found: ID does not exist" Oct 07 23:39:16 crc kubenswrapper[4871]: I1007 23:39:16.996336 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" path="/var/lib/kubelet/pods/2dbeb7fc-74b2-4dbf-be4e-ce2570c56365/volumes" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.044959 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-6jsb5"] Oct 07 23:39:24 crc kubenswrapper[4871]: E1007 23:39:24.045906 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerName="extract-utilities" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.045925 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerName="extract-utilities" Oct 07 23:39:24 crc kubenswrapper[4871]: E1007 23:39:24.045943 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerName="registry-server" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.045953 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerName="registry-server" Oct 07 23:39:24 crc kubenswrapper[4871]: E1007 23:39:24.045967 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerName="extract-content" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.045975 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerName="extract-content" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.046211 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dbeb7fc-74b2-4dbf-be4e-ce2570c56365" containerName="registry-server" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.046911 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6jsb5" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.052554 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-6jsb5"] Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.058581 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6x8j\" (UniqueName: \"kubernetes.io/projected/80322d0e-4017-4540-8ec5-98fa13ed2917-kube-api-access-j6x8j\") pod \"barbican-db-create-6jsb5\" (UID: \"80322d0e-4017-4540-8ec5-98fa13ed2917\") " pod="openstack/barbican-db-create-6jsb5" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.160878 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6x8j\" (UniqueName: \"kubernetes.io/projected/80322d0e-4017-4540-8ec5-98fa13ed2917-kube-api-access-j6x8j\") pod \"barbican-db-create-6jsb5\" (UID: \"80322d0e-4017-4540-8ec5-98fa13ed2917\") " pod="openstack/barbican-db-create-6jsb5" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.191535 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6x8j\" (UniqueName: \"kubernetes.io/projected/80322d0e-4017-4540-8ec5-98fa13ed2917-kube-api-access-j6x8j\") pod \"barbican-db-create-6jsb5\" (UID: \"80322d0e-4017-4540-8ec5-98fa13ed2917\") " pod="openstack/barbican-db-create-6jsb5" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.365698 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6jsb5" Oct 07 23:39:24 crc kubenswrapper[4871]: I1007 23:39:24.613827 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-6jsb5"] Oct 07 23:39:25 crc kubenswrapper[4871]: I1007 23:39:25.227175 4871 generic.go:334] "Generic (PLEG): container finished" podID="80322d0e-4017-4540-8ec5-98fa13ed2917" containerID="85a8da9f1c3210181112ade3024b0a769ed0e283f5389f18d35e8f5b80b0727f" exitCode=0 Oct 07 23:39:25 crc kubenswrapper[4871]: I1007 23:39:25.227270 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6jsb5" event={"ID":"80322d0e-4017-4540-8ec5-98fa13ed2917","Type":"ContainerDied","Data":"85a8da9f1c3210181112ade3024b0a769ed0e283f5389f18d35e8f5b80b0727f"} Oct 07 23:39:25 crc kubenswrapper[4871]: I1007 23:39:25.227596 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6jsb5" event={"ID":"80322d0e-4017-4540-8ec5-98fa13ed2917","Type":"ContainerStarted","Data":"1a1c579460d8318c772e07826b1a33843ed7a8085df15177c14ffc47e6f238f9"} Oct 07 23:39:26 crc kubenswrapper[4871]: I1007 23:39:26.560719 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6jsb5" Oct 07 23:39:26 crc kubenswrapper[4871]: I1007 23:39:26.605030 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6x8j\" (UniqueName: \"kubernetes.io/projected/80322d0e-4017-4540-8ec5-98fa13ed2917-kube-api-access-j6x8j\") pod \"80322d0e-4017-4540-8ec5-98fa13ed2917\" (UID: \"80322d0e-4017-4540-8ec5-98fa13ed2917\") " Oct 07 23:39:26 crc kubenswrapper[4871]: I1007 23:39:26.615871 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80322d0e-4017-4540-8ec5-98fa13ed2917-kube-api-access-j6x8j" (OuterVolumeSpecName: "kube-api-access-j6x8j") pod "80322d0e-4017-4540-8ec5-98fa13ed2917" (UID: "80322d0e-4017-4540-8ec5-98fa13ed2917"). InnerVolumeSpecName "kube-api-access-j6x8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:39:26 crc kubenswrapper[4871]: I1007 23:39:26.707188 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6x8j\" (UniqueName: \"kubernetes.io/projected/80322d0e-4017-4540-8ec5-98fa13ed2917-kube-api-access-j6x8j\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:27 crc kubenswrapper[4871]: I1007 23:39:27.266207 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6jsb5" event={"ID":"80322d0e-4017-4540-8ec5-98fa13ed2917","Type":"ContainerDied","Data":"1a1c579460d8318c772e07826b1a33843ed7a8085df15177c14ffc47e6f238f9"} Oct 07 23:39:27 crc kubenswrapper[4871]: I1007 23:39:27.266253 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a1c579460d8318c772e07826b1a33843ed7a8085df15177c14ffc47e6f238f9" Oct 07 23:39:27 crc kubenswrapper[4871]: I1007 23:39:27.266319 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6jsb5" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.079540 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-0c02-account-create-qhdtz"] Oct 07 23:39:34 crc kubenswrapper[4871]: E1007 23:39:34.080671 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80322d0e-4017-4540-8ec5-98fa13ed2917" containerName="mariadb-database-create" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.080690 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="80322d0e-4017-4540-8ec5-98fa13ed2917" containerName="mariadb-database-create" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.080967 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="80322d0e-4017-4540-8ec5-98fa13ed2917" containerName="mariadb-database-create" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.081614 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0c02-account-create-qhdtz" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.084380 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.101099 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0c02-account-create-qhdtz"] Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.143511 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5phw\" (UniqueName: \"kubernetes.io/projected/c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d-kube-api-access-r5phw\") pod \"barbican-0c02-account-create-qhdtz\" (UID: \"c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d\") " pod="openstack/barbican-0c02-account-create-qhdtz" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.246001 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5phw\" (UniqueName: \"kubernetes.io/projected/c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d-kube-api-access-r5phw\") pod \"barbican-0c02-account-create-qhdtz\" (UID: \"c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d\") " pod="openstack/barbican-0c02-account-create-qhdtz" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.271313 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5phw\" (UniqueName: \"kubernetes.io/projected/c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d-kube-api-access-r5phw\") pod \"barbican-0c02-account-create-qhdtz\" (UID: \"c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d\") " pod="openstack/barbican-0c02-account-create-qhdtz" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.419850 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0c02-account-create-qhdtz" Oct 07 23:39:34 crc kubenswrapper[4871]: I1007 23:39:34.689994 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0c02-account-create-qhdtz"] Oct 07 23:39:35 crc kubenswrapper[4871]: I1007 23:39:35.353516 4871 generic.go:334] "Generic (PLEG): container finished" podID="c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d" containerID="20765c19ad6595246dd2f00dadbd6ec9d43e91afd3aed34e564a56681fda5d61" exitCode=0 Oct 07 23:39:35 crc kubenswrapper[4871]: I1007 23:39:35.353645 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0c02-account-create-qhdtz" event={"ID":"c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d","Type":"ContainerDied","Data":"20765c19ad6595246dd2f00dadbd6ec9d43e91afd3aed34e564a56681fda5d61"} Oct 07 23:39:35 crc kubenswrapper[4871]: I1007 23:39:35.353976 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0c02-account-create-qhdtz" event={"ID":"c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d","Type":"ContainerStarted","Data":"0f0a00c71cc91d8ebbedacf42f0804e2f4fa073e51377664eb35434ceae1f71a"} Oct 07 23:39:35 crc kubenswrapper[4871]: I1007 23:39:35.512361 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:39:35 crc kubenswrapper[4871]: I1007 23:39:35.512442 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:39:36 crc kubenswrapper[4871]: I1007 23:39:36.737806 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0c02-account-create-qhdtz" Oct 07 23:39:36 crc kubenswrapper[4871]: I1007 23:39:36.790395 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5phw\" (UniqueName: \"kubernetes.io/projected/c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d-kube-api-access-r5phw\") pod \"c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d\" (UID: \"c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d\") " Oct 07 23:39:36 crc kubenswrapper[4871]: I1007 23:39:36.800226 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d-kube-api-access-r5phw" (OuterVolumeSpecName: "kube-api-access-r5phw") pod "c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d" (UID: "c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d"). InnerVolumeSpecName "kube-api-access-r5phw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:39:36 crc kubenswrapper[4871]: I1007 23:39:36.892693 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5phw\" (UniqueName: \"kubernetes.io/projected/c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d-kube-api-access-r5phw\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:37 crc kubenswrapper[4871]: I1007 23:39:37.384519 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0c02-account-create-qhdtz" event={"ID":"c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d","Type":"ContainerDied","Data":"0f0a00c71cc91d8ebbedacf42f0804e2f4fa073e51377664eb35434ceae1f71a"} Oct 07 23:39:37 crc kubenswrapper[4871]: I1007 23:39:37.385098 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f0a00c71cc91d8ebbedacf42f0804e2f4fa073e51377664eb35434ceae1f71a" Oct 07 23:39:37 crc kubenswrapper[4871]: I1007 23:39:37.384612 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0c02-account-create-qhdtz" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.319832 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dkp8g"] Oct 07 23:39:39 crc kubenswrapper[4871]: E1007 23:39:39.321321 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d" containerName="mariadb-account-create" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.321460 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d" containerName="mariadb-account-create" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.321987 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d" containerName="mariadb-account-create" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.324352 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.331107 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.331728 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-sssjq" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.335889 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dkp8g"] Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.444584 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4xql\" (UniqueName: \"kubernetes.io/projected/839e7e97-bf4c-4436-9fa5-9c45426422ff-kube-api-access-p4xql\") pod \"barbican-db-sync-dkp8g\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.444676 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-combined-ca-bundle\") pod \"barbican-db-sync-dkp8g\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.444981 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-db-sync-config-data\") pod \"barbican-db-sync-dkp8g\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.546941 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-db-sync-config-data\") pod \"barbican-db-sync-dkp8g\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.547374 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4xql\" (UniqueName: \"kubernetes.io/projected/839e7e97-bf4c-4436-9fa5-9c45426422ff-kube-api-access-p4xql\") pod \"barbican-db-sync-dkp8g\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.548144 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-combined-ca-bundle\") pod \"barbican-db-sync-dkp8g\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.552172 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-db-sync-config-data\") pod \"barbican-db-sync-dkp8g\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.553631 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-combined-ca-bundle\") pod \"barbican-db-sync-dkp8g\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.585746 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4xql\" (UniqueName: \"kubernetes.io/projected/839e7e97-bf4c-4436-9fa5-9c45426422ff-kube-api-access-p4xql\") pod \"barbican-db-sync-dkp8g\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:39 crc kubenswrapper[4871]: I1007 23:39:39.656228 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:40 crc kubenswrapper[4871]: I1007 23:39:40.219233 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dkp8g"] Oct 07 23:39:40 crc kubenswrapper[4871]: W1007 23:39:40.272683 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod839e7e97_bf4c_4436_9fa5_9c45426422ff.slice/crio-1b9f210a7799ab0bb1f93489fb986582f94b4e1e8b8a53a74714b236261b0e06 WatchSource:0}: Error finding container 1b9f210a7799ab0bb1f93489fb986582f94b4e1e8b8a53a74714b236261b0e06: Status 404 returned error can't find the container with id 1b9f210a7799ab0bb1f93489fb986582f94b4e1e8b8a53a74714b236261b0e06 Oct 07 23:39:40 crc kubenswrapper[4871]: I1007 23:39:40.410670 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dkp8g" event={"ID":"839e7e97-bf4c-4436-9fa5-9c45426422ff","Type":"ContainerStarted","Data":"1b9f210a7799ab0bb1f93489fb986582f94b4e1e8b8a53a74714b236261b0e06"} Oct 07 23:39:41 crc kubenswrapper[4871]: I1007 23:39:41.421830 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dkp8g" event={"ID":"839e7e97-bf4c-4436-9fa5-9c45426422ff","Type":"ContainerStarted","Data":"410417bc7820ae652758e1d45f46751f7247c7852fb25ca8ddd575ca2c2baf2d"} Oct 07 23:39:41 crc kubenswrapper[4871]: I1007 23:39:41.448221 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dkp8g" podStartSLOduration=2.448201489 podStartE2EDuration="2.448201489s" podCreationTimestamp="2025-10-07 23:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:39:41.440882985 +0000 UTC m=+5455.243581068" watchObservedRunningTime="2025-10-07 23:39:41.448201489 +0000 UTC m=+5455.250899562" Oct 07 23:39:42 crc kubenswrapper[4871]: I1007 23:39:42.435761 4871 generic.go:334] "Generic (PLEG): container finished" podID="839e7e97-bf4c-4436-9fa5-9c45426422ff" containerID="410417bc7820ae652758e1d45f46751f7247c7852fb25ca8ddd575ca2c2baf2d" exitCode=0 Oct 07 23:39:42 crc kubenswrapper[4871]: I1007 23:39:42.435866 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dkp8g" event={"ID":"839e7e97-bf4c-4436-9fa5-9c45426422ff","Type":"ContainerDied","Data":"410417bc7820ae652758e1d45f46751f7247c7852fb25ca8ddd575ca2c2baf2d"} Oct 07 23:39:43 crc kubenswrapper[4871]: I1007 23:39:43.812773 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:43 crc kubenswrapper[4871]: I1007 23:39:43.922443 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-db-sync-config-data\") pod \"839e7e97-bf4c-4436-9fa5-9c45426422ff\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " Oct 07 23:39:43 crc kubenswrapper[4871]: I1007 23:39:43.922664 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4xql\" (UniqueName: \"kubernetes.io/projected/839e7e97-bf4c-4436-9fa5-9c45426422ff-kube-api-access-p4xql\") pod \"839e7e97-bf4c-4436-9fa5-9c45426422ff\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " Oct 07 23:39:43 crc kubenswrapper[4871]: I1007 23:39:43.922898 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-combined-ca-bundle\") pod \"839e7e97-bf4c-4436-9fa5-9c45426422ff\" (UID: \"839e7e97-bf4c-4436-9fa5-9c45426422ff\") " Oct 07 23:39:43 crc kubenswrapper[4871]: I1007 23:39:43.927717 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/839e7e97-bf4c-4436-9fa5-9c45426422ff-kube-api-access-p4xql" (OuterVolumeSpecName: "kube-api-access-p4xql") pod "839e7e97-bf4c-4436-9fa5-9c45426422ff" (UID: "839e7e97-bf4c-4436-9fa5-9c45426422ff"). InnerVolumeSpecName "kube-api-access-p4xql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:39:43 crc kubenswrapper[4871]: I1007 23:39:43.929621 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "839e7e97-bf4c-4436-9fa5-9c45426422ff" (UID: "839e7e97-bf4c-4436-9fa5-9c45426422ff"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:39:43 crc kubenswrapper[4871]: I1007 23:39:43.964526 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "839e7e97-bf4c-4436-9fa5-9c45426422ff" (UID: "839e7e97-bf4c-4436-9fa5-9c45426422ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.024504 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.024528 4871 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/839e7e97-bf4c-4436-9fa5-9c45426422ff-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.024538 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4xql\" (UniqueName: \"kubernetes.io/projected/839e7e97-bf4c-4436-9fa5-9c45426422ff-kube-api-access-p4xql\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.487930 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dkp8g" event={"ID":"839e7e97-bf4c-4436-9fa5-9c45426422ff","Type":"ContainerDied","Data":"1b9f210a7799ab0bb1f93489fb986582f94b4e1e8b8a53a74714b236261b0e06"} Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.488309 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b9f210a7799ab0bb1f93489fb986582f94b4e1e8b8a53a74714b236261b0e06" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.487998 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dkp8g" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.748990 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-74dc8ccf74-5tktj"] Oct 07 23:39:44 crc kubenswrapper[4871]: E1007 23:39:44.749402 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="839e7e97-bf4c-4436-9fa5-9c45426422ff" containerName="barbican-db-sync" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.749425 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="839e7e97-bf4c-4436-9fa5-9c45426422ff" containerName="barbican-db-sync" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.749653 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="839e7e97-bf4c-4436-9fa5-9c45426422ff" containerName="barbican-db-sync" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.750745 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.753349 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.753534 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-sssjq" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.753682 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.774742 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5cc676b4f5-92cx4"] Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.776420 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.779208 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.784619 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-74dc8ccf74-5tktj"] Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.794875 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5cc676b4f5-92cx4"] Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.848635 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d44dfd697-wnr8b"] Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.849955 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.863969 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d44dfd697-wnr8b"] Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941396 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdcrm\" (UniqueName: \"kubernetes.io/projected/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-kube-api-access-xdcrm\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941466 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-logs\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941484 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-combined-ca-bundle\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941508 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-config-data-custom\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941529 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-logs\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941556 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcjbj\" (UniqueName: \"kubernetes.io/projected/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-kube-api-access-lcjbj\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941616 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-config-data\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941641 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-config-data-custom\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941658 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-config-data\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.941683 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-combined-ca-bundle\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.947844 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6849865b9b-99rd5"] Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.949267 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.950914 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 07 23:39:44 crc kubenswrapper[4871]: I1007 23:39:44.953812 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6849865b9b-99rd5"] Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043119 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-config-data-custom\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043164 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-config-data\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043196 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61ba6fb5-5888-410b-938a-dc06179dd26f-logs\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043220 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-combined-ca-bundle\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043239 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61ba6fb5-5888-410b-938a-dc06179dd26f-combined-ca-bundle\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043265 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5zpz\" (UniqueName: \"kubernetes.io/projected/5548ed74-08b9-42c3-8d2c-99e82f860a92-kube-api-access-q5zpz\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043295 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-sb\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043314 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdcrm\" (UniqueName: \"kubernetes.io/projected/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-kube-api-access-xdcrm\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043528 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-logs\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043583 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-combined-ca-bundle\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043645 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-config-data-custom\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043690 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-logs\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043749 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rkzs\" (UniqueName: \"kubernetes.io/projected/61ba6fb5-5888-410b-938a-dc06179dd26f-kube-api-access-9rkzs\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043780 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61ba6fb5-5888-410b-938a-dc06179dd26f-config-data-custom\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043834 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcjbj\" (UniqueName: \"kubernetes.io/projected/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-kube-api-access-lcjbj\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043868 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-dns-svc\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043912 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61ba6fb5-5888-410b-938a-dc06179dd26f-config-data\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.043994 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-nb\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.044023 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-config\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.044084 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-logs\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.044108 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-config-data\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.044437 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-logs\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.047756 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-config-data-custom\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.047909 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-combined-ca-bundle\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.048657 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-config-data-custom\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.049516 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-config-data\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.054916 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-config-data\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.059612 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-combined-ca-bundle\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.059648 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdcrm\" (UniqueName: \"kubernetes.io/projected/e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7-kube-api-access-xdcrm\") pod \"barbican-keystone-listener-74dc8ccf74-5tktj\" (UID: \"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7\") " pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.062059 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcjbj\" (UniqueName: \"kubernetes.io/projected/f2cf443f-a41e-4c1a-a488-0a90f44b8e74-kube-api-access-lcjbj\") pod \"barbican-worker-5cc676b4f5-92cx4\" (UID: \"f2cf443f-a41e-4c1a-a488-0a90f44b8e74\") " pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.072474 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.096041 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5cc676b4f5-92cx4" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145366 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61ba6fb5-5888-410b-938a-dc06179dd26f-logs\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145420 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61ba6fb5-5888-410b-938a-dc06179dd26f-combined-ca-bundle\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145448 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5zpz\" (UniqueName: \"kubernetes.io/projected/5548ed74-08b9-42c3-8d2c-99e82f860a92-kube-api-access-q5zpz\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145481 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-sb\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145542 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rkzs\" (UniqueName: \"kubernetes.io/projected/61ba6fb5-5888-410b-938a-dc06179dd26f-kube-api-access-9rkzs\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145563 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61ba6fb5-5888-410b-938a-dc06179dd26f-config-data-custom\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145585 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-dns-svc\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145607 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61ba6fb5-5888-410b-938a-dc06179dd26f-config-data\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145634 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-nb\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145650 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-config\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.145928 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61ba6fb5-5888-410b-938a-dc06179dd26f-logs\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.146625 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-config\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.146631 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-nb\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.146991 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-sb\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.147342 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-dns-svc\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.149540 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/61ba6fb5-5888-410b-938a-dc06179dd26f-config-data-custom\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.150302 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61ba6fb5-5888-410b-938a-dc06179dd26f-combined-ca-bundle\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.163147 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61ba6fb5-5888-410b-938a-dc06179dd26f-config-data\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.170071 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5zpz\" (UniqueName: \"kubernetes.io/projected/5548ed74-08b9-42c3-8d2c-99e82f860a92-kube-api-access-q5zpz\") pod \"dnsmasq-dns-d44dfd697-wnr8b\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.177187 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rkzs\" (UniqueName: \"kubernetes.io/projected/61ba6fb5-5888-410b-938a-dc06179dd26f-kube-api-access-9rkzs\") pod \"barbican-api-6849865b9b-99rd5\" (UID: \"61ba6fb5-5888-410b-938a-dc06179dd26f\") " pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.185899 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.270994 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.565732 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-74dc8ccf74-5tktj"] Oct 07 23:39:45 crc kubenswrapper[4871]: W1007 23:39:45.643147 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2cf443f_a41e_4c1a_a488_0a90f44b8e74.slice/crio-0f4132db48b2f942a129850a8bc25e7ebc648a2b70815419c4cda5e170845805 WatchSource:0}: Error finding container 0f4132db48b2f942a129850a8bc25e7ebc648a2b70815419c4cda5e170845805: Status 404 returned error can't find the container with id 0f4132db48b2f942a129850a8bc25e7ebc648a2b70815419c4cda5e170845805 Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.646651 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5cc676b4f5-92cx4"] Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.714637 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d44dfd697-wnr8b"] Oct 07 23:39:45 crc kubenswrapper[4871]: I1007 23:39:45.782228 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6849865b9b-99rd5"] Oct 07 23:39:45 crc kubenswrapper[4871]: W1007 23:39:45.796747 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61ba6fb5_5888_410b_938a_dc06179dd26f.slice/crio-b7f865a6def9ef0ee66644afcf3bcae29903d8d7d1d0353cca1ea3cdad6f1a59 WatchSource:0}: Error finding container b7f865a6def9ef0ee66644afcf3bcae29903d8d7d1d0353cca1ea3cdad6f1a59: Status 404 returned error can't find the container with id b7f865a6def9ef0ee66644afcf3bcae29903d8d7d1d0353cca1ea3cdad6f1a59 Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.503996 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5cc676b4f5-92cx4" event={"ID":"f2cf443f-a41e-4c1a-a488-0a90f44b8e74","Type":"ContainerStarted","Data":"728c847f7310dc3e0d71ab919b030aa250b37aafa41c9da42e1fa27b862121c7"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.504381 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5cc676b4f5-92cx4" event={"ID":"f2cf443f-a41e-4c1a-a488-0a90f44b8e74","Type":"ContainerStarted","Data":"41ede0442d2bd3ffec264e462b92a24f6efe27088e3e799278c878af0e38f090"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.504397 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5cc676b4f5-92cx4" event={"ID":"f2cf443f-a41e-4c1a-a488-0a90f44b8e74","Type":"ContainerStarted","Data":"0f4132db48b2f942a129850a8bc25e7ebc648a2b70815419c4cda5e170845805"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.507367 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6849865b9b-99rd5" event={"ID":"61ba6fb5-5888-410b-938a-dc06179dd26f","Type":"ContainerStarted","Data":"d8d7a318810f9c7b06577cd4235a62f765f48f4ee1c244e212ca6850c02b8165"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.507401 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6849865b9b-99rd5" event={"ID":"61ba6fb5-5888-410b-938a-dc06179dd26f","Type":"ContainerStarted","Data":"bf9097bfd0633dbf57aced915618fb5a7a767d7874eecfd4a63aa14976432567"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.507410 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6849865b9b-99rd5" event={"ID":"61ba6fb5-5888-410b-938a-dc06179dd26f","Type":"ContainerStarted","Data":"b7f865a6def9ef0ee66644afcf3bcae29903d8d7d1d0353cca1ea3cdad6f1a59"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.507538 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.509380 4871 generic.go:334] "Generic (PLEG): container finished" podID="5548ed74-08b9-42c3-8d2c-99e82f860a92" containerID="f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503" exitCode=0 Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.509424 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" event={"ID":"5548ed74-08b9-42c3-8d2c-99e82f860a92","Type":"ContainerDied","Data":"f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.509455 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" event={"ID":"5548ed74-08b9-42c3-8d2c-99e82f860a92","Type":"ContainerStarted","Data":"9988ce2323a94fa1876ce8febd438555517d419fec5bd7d09c73e50a16345974"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.513411 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" event={"ID":"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7","Type":"ContainerStarted","Data":"a202bbd8a2bdf74ad384e915099b390501994696c4dffc9bd22cafeff76d79a2"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.513453 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" event={"ID":"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7","Type":"ContainerStarted","Data":"bce0c8cafbc226dc33ba4d043af07e2ef3d9a610f71fa4a6169ec5d851430042"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.513467 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" event={"ID":"e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7","Type":"ContainerStarted","Data":"453e90c59290c851d80f4fa0e3bf58843459cdeeb8f5677535b3798c5f09e482"} Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.530289 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5cc676b4f5-92cx4" podStartSLOduration=2.5302685609999997 podStartE2EDuration="2.530268561s" podCreationTimestamp="2025-10-07 23:39:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:39:46.529216413 +0000 UTC m=+5460.331914526" watchObservedRunningTime="2025-10-07 23:39:46.530268561 +0000 UTC m=+5460.332966634" Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.562451 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6849865b9b-99rd5" podStartSLOduration=2.562416979 podStartE2EDuration="2.562416979s" podCreationTimestamp="2025-10-07 23:39:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:39:46.549133709 +0000 UTC m=+5460.351831772" watchObservedRunningTime="2025-10-07 23:39:46.562416979 +0000 UTC m=+5460.365115052" Oct 07 23:39:46 crc kubenswrapper[4871]: I1007 23:39:46.596361 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-74dc8ccf74-5tktj" podStartSLOduration=2.596336045 podStartE2EDuration="2.596336045s" podCreationTimestamp="2025-10-07 23:39:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:39:46.572090995 +0000 UTC m=+5460.374789068" watchObservedRunningTime="2025-10-07 23:39:46.596336045 +0000 UTC m=+5460.399034108" Oct 07 23:39:47 crc kubenswrapper[4871]: I1007 23:39:47.526248 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" event={"ID":"5548ed74-08b9-42c3-8d2c-99e82f860a92","Type":"ContainerStarted","Data":"2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f"} Oct 07 23:39:47 crc kubenswrapper[4871]: I1007 23:39:47.526835 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:48 crc kubenswrapper[4871]: I1007 23:39:48.534655 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:51 crc kubenswrapper[4871]: I1007 23:39:51.732502 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:51 crc kubenswrapper[4871]: I1007 23:39:51.769745 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" podStartSLOduration=7.769712008 podStartE2EDuration="7.769712008s" podCreationTimestamp="2025-10-07 23:39:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:39:47.55337349 +0000 UTC m=+5461.356071613" watchObservedRunningTime="2025-10-07 23:39:51.769712008 +0000 UTC m=+5465.572410121" Oct 07 23:39:53 crc kubenswrapper[4871]: I1007 23:39:53.120560 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6849865b9b-99rd5" Oct 07 23:39:55 crc kubenswrapper[4871]: I1007 23:39:55.187429 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:39:55 crc kubenswrapper[4871]: I1007 23:39:55.246719 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f5d545f5f-gps7v"] Oct 07 23:39:55 crc kubenswrapper[4871]: I1007 23:39:55.247226 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" podUID="98667f31-9a8f-478f-9be8-227e60b93ced" containerName="dnsmasq-dns" containerID="cri-o://e63ab012562588b6176ccfa30e05ce55083313fd4834e5b265207c858bcbe2f4" gracePeriod=10 Oct 07 23:39:55 crc kubenswrapper[4871]: I1007 23:39:55.607994 4871 generic.go:334] "Generic (PLEG): container finished" podID="98667f31-9a8f-478f-9be8-227e60b93ced" containerID="e63ab012562588b6176ccfa30e05ce55083313fd4834e5b265207c858bcbe2f4" exitCode=0 Oct 07 23:39:55 crc kubenswrapper[4871]: I1007 23:39:55.608045 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" event={"ID":"98667f31-9a8f-478f-9be8-227e60b93ced","Type":"ContainerDied","Data":"e63ab012562588b6176ccfa30e05ce55083313fd4834e5b265207c858bcbe2f4"} Oct 07 23:39:55 crc kubenswrapper[4871]: I1007 23:39:55.870610 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.049482 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-nb\") pod \"98667f31-9a8f-478f-9be8-227e60b93ced\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.049524 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-sb\") pod \"98667f31-9a8f-478f-9be8-227e60b93ced\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.049618 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-config\") pod \"98667f31-9a8f-478f-9be8-227e60b93ced\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.049672 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4fsz\" (UniqueName: \"kubernetes.io/projected/98667f31-9a8f-478f-9be8-227e60b93ced-kube-api-access-n4fsz\") pod \"98667f31-9a8f-478f-9be8-227e60b93ced\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.049700 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-dns-svc\") pod \"98667f31-9a8f-478f-9be8-227e60b93ced\" (UID: \"98667f31-9a8f-478f-9be8-227e60b93ced\") " Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.056043 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98667f31-9a8f-478f-9be8-227e60b93ced-kube-api-access-n4fsz" (OuterVolumeSpecName: "kube-api-access-n4fsz") pod "98667f31-9a8f-478f-9be8-227e60b93ced" (UID: "98667f31-9a8f-478f-9be8-227e60b93ced"). InnerVolumeSpecName "kube-api-access-n4fsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.093171 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "98667f31-9a8f-478f-9be8-227e60b93ced" (UID: "98667f31-9a8f-478f-9be8-227e60b93ced"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.099084 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "98667f31-9a8f-478f-9be8-227e60b93ced" (UID: "98667f31-9a8f-478f-9be8-227e60b93ced"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.101304 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-config" (OuterVolumeSpecName: "config") pod "98667f31-9a8f-478f-9be8-227e60b93ced" (UID: "98667f31-9a8f-478f-9be8-227e60b93ced"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.115229 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "98667f31-9a8f-478f-9be8-227e60b93ced" (UID: "98667f31-9a8f-478f-9be8-227e60b93ced"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.151667 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.151696 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.151706 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.151716 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4fsz\" (UniqueName: \"kubernetes.io/projected/98667f31-9a8f-478f-9be8-227e60b93ced-kube-api-access-n4fsz\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.151725 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98667f31-9a8f-478f-9be8-227e60b93ced-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.619246 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" event={"ID":"98667f31-9a8f-478f-9be8-227e60b93ced","Type":"ContainerDied","Data":"43759222c9f8133b25544bc5de41d772bff808df637960d721068dc1e5d49f9d"} Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.619296 4871 scope.go:117] "RemoveContainer" containerID="e63ab012562588b6176ccfa30e05ce55083313fd4834e5b265207c858bcbe2f4" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.619407 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f5d545f5f-gps7v" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.661714 4871 scope.go:117] "RemoveContainer" containerID="3253bf3df816aa3542e3c3d3afed8e44a89919ab2e13c4cf8ba2ed863e0cb2f7" Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.667883 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f5d545f5f-gps7v"] Oct 07 23:39:56 crc kubenswrapper[4871]: I1007 23:39:56.675620 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f5d545f5f-gps7v"] Oct 07 23:39:57 crc kubenswrapper[4871]: I1007 23:39:57.002979 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98667f31-9a8f-478f-9be8-227e60b93ced" path="/var/lib/kubelet/pods/98667f31-9a8f-478f-9be8-227e60b93ced/volumes" Oct 07 23:40:05 crc kubenswrapper[4871]: I1007 23:40:05.511990 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:40:05 crc kubenswrapper[4871]: I1007 23:40:05.512518 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.091149 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-xln4c"] Oct 07 23:40:08 crc kubenswrapper[4871]: E1007 23:40:08.091741 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98667f31-9a8f-478f-9be8-227e60b93ced" containerName="init" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.091755 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="98667f31-9a8f-478f-9be8-227e60b93ced" containerName="init" Oct 07 23:40:08 crc kubenswrapper[4871]: E1007 23:40:08.091766 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98667f31-9a8f-478f-9be8-227e60b93ced" containerName="dnsmasq-dns" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.091774 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="98667f31-9a8f-478f-9be8-227e60b93ced" containerName="dnsmasq-dns" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.091972 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="98667f31-9a8f-478f-9be8-227e60b93ced" containerName="dnsmasq-dns" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.097375 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xln4c" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.113171 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xln4c"] Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.176527 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qwtj\" (UniqueName: \"kubernetes.io/projected/50de0223-7b4d-4035-818d-1d238a83dd89-kube-api-access-2qwtj\") pod \"neutron-db-create-xln4c\" (UID: \"50de0223-7b4d-4035-818d-1d238a83dd89\") " pod="openstack/neutron-db-create-xln4c" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.278388 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qwtj\" (UniqueName: \"kubernetes.io/projected/50de0223-7b4d-4035-818d-1d238a83dd89-kube-api-access-2qwtj\") pod \"neutron-db-create-xln4c\" (UID: \"50de0223-7b4d-4035-818d-1d238a83dd89\") " pod="openstack/neutron-db-create-xln4c" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.303747 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qwtj\" (UniqueName: \"kubernetes.io/projected/50de0223-7b4d-4035-818d-1d238a83dd89-kube-api-access-2qwtj\") pod \"neutron-db-create-xln4c\" (UID: \"50de0223-7b4d-4035-818d-1d238a83dd89\") " pod="openstack/neutron-db-create-xln4c" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.427117 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xln4c" Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.704356 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xln4c"] Oct 07 23:40:08 crc kubenswrapper[4871]: W1007 23:40:08.711553 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50de0223_7b4d_4035_818d_1d238a83dd89.slice/crio-881d50b63d171d30a8b3a6ff97546a1b96f0b19c5d327e22b7aaeeef23a614e5 WatchSource:0}: Error finding container 881d50b63d171d30a8b3a6ff97546a1b96f0b19c5d327e22b7aaeeef23a614e5: Status 404 returned error can't find the container with id 881d50b63d171d30a8b3a6ff97546a1b96f0b19c5d327e22b7aaeeef23a614e5 Oct 07 23:40:08 crc kubenswrapper[4871]: I1007 23:40:08.749870 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xln4c" event={"ID":"50de0223-7b4d-4035-818d-1d238a83dd89","Type":"ContainerStarted","Data":"881d50b63d171d30a8b3a6ff97546a1b96f0b19c5d327e22b7aaeeef23a614e5"} Oct 07 23:40:09 crc kubenswrapper[4871]: I1007 23:40:09.761737 4871 generic.go:334] "Generic (PLEG): container finished" podID="50de0223-7b4d-4035-818d-1d238a83dd89" containerID="6de56f1138da3a53c6e3bff9654a0471c09c87d61fd117357e77f6098e352504" exitCode=0 Oct 07 23:40:09 crc kubenswrapper[4871]: I1007 23:40:09.762088 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xln4c" event={"ID":"50de0223-7b4d-4035-818d-1d238a83dd89","Type":"ContainerDied","Data":"6de56f1138da3a53c6e3bff9654a0471c09c87d61fd117357e77f6098e352504"} Oct 07 23:40:11 crc kubenswrapper[4871]: I1007 23:40:11.194345 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xln4c" Oct 07 23:40:11 crc kubenswrapper[4871]: I1007 23:40:11.334526 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qwtj\" (UniqueName: \"kubernetes.io/projected/50de0223-7b4d-4035-818d-1d238a83dd89-kube-api-access-2qwtj\") pod \"50de0223-7b4d-4035-818d-1d238a83dd89\" (UID: \"50de0223-7b4d-4035-818d-1d238a83dd89\") " Oct 07 23:40:11 crc kubenswrapper[4871]: I1007 23:40:11.340589 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50de0223-7b4d-4035-818d-1d238a83dd89-kube-api-access-2qwtj" (OuterVolumeSpecName: "kube-api-access-2qwtj") pod "50de0223-7b4d-4035-818d-1d238a83dd89" (UID: "50de0223-7b4d-4035-818d-1d238a83dd89"). InnerVolumeSpecName "kube-api-access-2qwtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:40:11 crc kubenswrapper[4871]: I1007 23:40:11.436319 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qwtj\" (UniqueName: \"kubernetes.io/projected/50de0223-7b4d-4035-818d-1d238a83dd89-kube-api-access-2qwtj\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:11 crc kubenswrapper[4871]: I1007 23:40:11.783084 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xln4c" event={"ID":"50de0223-7b4d-4035-818d-1d238a83dd89","Type":"ContainerDied","Data":"881d50b63d171d30a8b3a6ff97546a1b96f0b19c5d327e22b7aaeeef23a614e5"} Oct 07 23:40:11 crc kubenswrapper[4871]: I1007 23:40:11.783126 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="881d50b63d171d30a8b3a6ff97546a1b96f0b19c5d327e22b7aaeeef23a614e5" Oct 07 23:40:11 crc kubenswrapper[4871]: I1007 23:40:11.783194 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xln4c" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.138508 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2101-account-create-f8565"] Oct 07 23:40:18 crc kubenswrapper[4871]: E1007 23:40:18.140058 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50de0223-7b4d-4035-818d-1d238a83dd89" containerName="mariadb-database-create" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.140092 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="50de0223-7b4d-4035-818d-1d238a83dd89" containerName="mariadb-database-create" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.140605 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="50de0223-7b4d-4035-818d-1d238a83dd89" containerName="mariadb-database-create" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.142012 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2101-account-create-f8565" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.144954 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.158688 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2101-account-create-f8565"] Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.260932 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw7hq\" (UniqueName: \"kubernetes.io/projected/99101738-90d6-4b8a-98b3-10e9af17c641-kube-api-access-dw7hq\") pod \"neutron-2101-account-create-f8565\" (UID: \"99101738-90d6-4b8a-98b3-10e9af17c641\") " pod="openstack/neutron-2101-account-create-f8565" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.362400 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw7hq\" (UniqueName: \"kubernetes.io/projected/99101738-90d6-4b8a-98b3-10e9af17c641-kube-api-access-dw7hq\") pod \"neutron-2101-account-create-f8565\" (UID: \"99101738-90d6-4b8a-98b3-10e9af17c641\") " pod="openstack/neutron-2101-account-create-f8565" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.383159 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw7hq\" (UniqueName: \"kubernetes.io/projected/99101738-90d6-4b8a-98b3-10e9af17c641-kube-api-access-dw7hq\") pod \"neutron-2101-account-create-f8565\" (UID: \"99101738-90d6-4b8a-98b3-10e9af17c641\") " pod="openstack/neutron-2101-account-create-f8565" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.470422 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2101-account-create-f8565" Oct 07 23:40:18 crc kubenswrapper[4871]: I1007 23:40:18.913591 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2101-account-create-f8565"] Oct 07 23:40:19 crc kubenswrapper[4871]: I1007 23:40:19.868223 4871 generic.go:334] "Generic (PLEG): container finished" podID="99101738-90d6-4b8a-98b3-10e9af17c641" containerID="daa5ddfe41d3c442593ac170da5ddcd02661e3fcfa7e169934d7c3ddfa038f42" exitCode=0 Oct 07 23:40:19 crc kubenswrapper[4871]: I1007 23:40:19.868318 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2101-account-create-f8565" event={"ID":"99101738-90d6-4b8a-98b3-10e9af17c641","Type":"ContainerDied","Data":"daa5ddfe41d3c442593ac170da5ddcd02661e3fcfa7e169934d7c3ddfa038f42"} Oct 07 23:40:19 crc kubenswrapper[4871]: I1007 23:40:19.868564 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2101-account-create-f8565" event={"ID":"99101738-90d6-4b8a-98b3-10e9af17c641","Type":"ContainerStarted","Data":"e009fc1859d9393117342357aebe8ed8bf4c992ca7704dc8f48f26ce3cc3a411"} Oct 07 23:40:21 crc kubenswrapper[4871]: I1007 23:40:21.305403 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2101-account-create-f8565" Oct 07 23:40:21 crc kubenswrapper[4871]: I1007 23:40:21.422380 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw7hq\" (UniqueName: \"kubernetes.io/projected/99101738-90d6-4b8a-98b3-10e9af17c641-kube-api-access-dw7hq\") pod \"99101738-90d6-4b8a-98b3-10e9af17c641\" (UID: \"99101738-90d6-4b8a-98b3-10e9af17c641\") " Oct 07 23:40:21 crc kubenswrapper[4871]: I1007 23:40:21.430057 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99101738-90d6-4b8a-98b3-10e9af17c641-kube-api-access-dw7hq" (OuterVolumeSpecName: "kube-api-access-dw7hq") pod "99101738-90d6-4b8a-98b3-10e9af17c641" (UID: "99101738-90d6-4b8a-98b3-10e9af17c641"). InnerVolumeSpecName "kube-api-access-dw7hq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:40:21 crc kubenswrapper[4871]: I1007 23:40:21.524068 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw7hq\" (UniqueName: \"kubernetes.io/projected/99101738-90d6-4b8a-98b3-10e9af17c641-kube-api-access-dw7hq\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:21 crc kubenswrapper[4871]: I1007 23:40:21.894746 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2101-account-create-f8565" event={"ID":"99101738-90d6-4b8a-98b3-10e9af17c641","Type":"ContainerDied","Data":"e009fc1859d9393117342357aebe8ed8bf4c992ca7704dc8f48f26ce3cc3a411"} Oct 07 23:40:21 crc kubenswrapper[4871]: I1007 23:40:21.894851 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e009fc1859d9393117342357aebe8ed8bf4c992ca7704dc8f48f26ce3cc3a411" Oct 07 23:40:21 crc kubenswrapper[4871]: I1007 23:40:21.894940 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2101-account-create-f8565" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.336970 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kf9sd"] Oct 07 23:40:23 crc kubenswrapper[4871]: E1007 23:40:23.337888 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99101738-90d6-4b8a-98b3-10e9af17c641" containerName="mariadb-account-create" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.337912 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="99101738-90d6-4b8a-98b3-10e9af17c641" containerName="mariadb-account-create" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.338191 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="99101738-90d6-4b8a-98b3-10e9af17c641" containerName="mariadb-account-create" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.338901 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.341689 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hgwbh" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.342009 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.342376 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.354868 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kf9sd"] Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.466911 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-combined-ca-bundle\") pod \"neutron-db-sync-kf9sd\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.467297 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbhcg\" (UniqueName: \"kubernetes.io/projected/45faaa8f-9462-4793-90ab-b22c394e7177-kube-api-access-dbhcg\") pod \"neutron-db-sync-kf9sd\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.467459 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-config\") pod \"neutron-db-sync-kf9sd\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.568941 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-combined-ca-bundle\") pod \"neutron-db-sync-kf9sd\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.569264 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbhcg\" (UniqueName: \"kubernetes.io/projected/45faaa8f-9462-4793-90ab-b22c394e7177-kube-api-access-dbhcg\") pod \"neutron-db-sync-kf9sd\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.569380 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-config\") pod \"neutron-db-sync-kf9sd\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.574646 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-config\") pod \"neutron-db-sync-kf9sd\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.574817 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-combined-ca-bundle\") pod \"neutron-db-sync-kf9sd\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.587682 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbhcg\" (UniqueName: \"kubernetes.io/projected/45faaa8f-9462-4793-90ab-b22c394e7177-kube-api-access-dbhcg\") pod \"neutron-db-sync-kf9sd\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:23 crc kubenswrapper[4871]: I1007 23:40:23.673569 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:24 crc kubenswrapper[4871]: I1007 23:40:24.137468 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kf9sd"] Oct 07 23:40:24 crc kubenswrapper[4871]: I1007 23:40:24.946277 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kf9sd" event={"ID":"45faaa8f-9462-4793-90ab-b22c394e7177","Type":"ContainerStarted","Data":"eb074beb4c7ef7b670094238e47500474752585e6a2999234668260465f87d8b"} Oct 07 23:40:24 crc kubenswrapper[4871]: I1007 23:40:24.946729 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kf9sd" event={"ID":"45faaa8f-9462-4793-90ab-b22c394e7177","Type":"ContainerStarted","Data":"1a7dad65e6254051ea9ed5bb588ca988b6a472397374cb183a5802975cb87bd1"} Oct 07 23:40:28 crc kubenswrapper[4871]: I1007 23:40:28.989874 4871 generic.go:334] "Generic (PLEG): container finished" podID="45faaa8f-9462-4793-90ab-b22c394e7177" containerID="eb074beb4c7ef7b670094238e47500474752585e6a2999234668260465f87d8b" exitCode=0 Oct 07 23:40:29 crc kubenswrapper[4871]: I1007 23:40:29.003316 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kf9sd" event={"ID":"45faaa8f-9462-4793-90ab-b22c394e7177","Type":"ContainerDied","Data":"eb074beb4c7ef7b670094238e47500474752585e6a2999234668260465f87d8b"} Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.298846 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.401249 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-config\") pod \"45faaa8f-9462-4793-90ab-b22c394e7177\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.401290 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-combined-ca-bundle\") pod \"45faaa8f-9462-4793-90ab-b22c394e7177\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.401450 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbhcg\" (UniqueName: \"kubernetes.io/projected/45faaa8f-9462-4793-90ab-b22c394e7177-kube-api-access-dbhcg\") pod \"45faaa8f-9462-4793-90ab-b22c394e7177\" (UID: \"45faaa8f-9462-4793-90ab-b22c394e7177\") " Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.407133 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45faaa8f-9462-4793-90ab-b22c394e7177-kube-api-access-dbhcg" (OuterVolumeSpecName: "kube-api-access-dbhcg") pod "45faaa8f-9462-4793-90ab-b22c394e7177" (UID: "45faaa8f-9462-4793-90ab-b22c394e7177"). InnerVolumeSpecName "kube-api-access-dbhcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.428394 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45faaa8f-9462-4793-90ab-b22c394e7177" (UID: "45faaa8f-9462-4793-90ab-b22c394e7177"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.430195 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-config" (OuterVolumeSpecName: "config") pod "45faaa8f-9462-4793-90ab-b22c394e7177" (UID: "45faaa8f-9462-4793-90ab-b22c394e7177"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.503145 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbhcg\" (UniqueName: \"kubernetes.io/projected/45faaa8f-9462-4793-90ab-b22c394e7177-kube-api-access-dbhcg\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.503168 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:30 crc kubenswrapper[4871]: I1007 23:40:30.503180 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45faaa8f-9462-4793-90ab-b22c394e7177-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.016449 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kf9sd" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.024777 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kf9sd" event={"ID":"45faaa8f-9462-4793-90ab-b22c394e7177","Type":"ContainerDied","Data":"1a7dad65e6254051ea9ed5bb588ca988b6a472397374cb183a5802975cb87bd1"} Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.024899 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a7dad65e6254051ea9ed5bb588ca988b6a472397374cb183a5802975cb87bd1" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.185615 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b7d74c6cf-5w7gm"] Oct 07 23:40:31 crc kubenswrapper[4871]: E1007 23:40:31.186438 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45faaa8f-9462-4793-90ab-b22c394e7177" containerName="neutron-db-sync" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.186468 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="45faaa8f-9462-4793-90ab-b22c394e7177" containerName="neutron-db-sync" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.186927 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="45faaa8f-9462-4793-90ab-b22c394e7177" containerName="neutron-db-sync" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.188564 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.206156 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b7d74c6cf-5w7gm"] Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.322887 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-65f4496dcc-6mkfj"] Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.324340 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.330724 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.330946 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hgwbh" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.331245 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.334194 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65f4496dcc-6mkfj"] Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.335699 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wshq5\" (UniqueName: \"kubernetes.io/projected/e228a39c-1929-416e-a1ab-15c66c6eb329-kube-api-access-wshq5\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.335733 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-nb\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.335782 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-sb\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.335886 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-dns-svc\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.335977 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-config\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.436876 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f242653b-4449-452c-a438-2fbdd2a87dba-config\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.437020 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-config\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.437149 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64ssc\" (UniqueName: \"kubernetes.io/projected/f242653b-4449-452c-a438-2fbdd2a87dba-kube-api-access-64ssc\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.437229 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wshq5\" (UniqueName: \"kubernetes.io/projected/e228a39c-1929-416e-a1ab-15c66c6eb329-kube-api-access-wshq5\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.437257 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-nb\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.437342 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-sb\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.437423 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-dns-svc\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.437448 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f242653b-4449-452c-a438-2fbdd2a87dba-combined-ca-bundle\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.437490 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f242653b-4449-452c-a438-2fbdd2a87dba-httpd-config\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.437877 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-config\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.438307 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-nb\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.438461 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-sb\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.438502 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-dns-svc\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.458889 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wshq5\" (UniqueName: \"kubernetes.io/projected/e228a39c-1929-416e-a1ab-15c66c6eb329-kube-api-access-wshq5\") pod \"dnsmasq-dns-7b7d74c6cf-5w7gm\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.507569 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.539599 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64ssc\" (UniqueName: \"kubernetes.io/projected/f242653b-4449-452c-a438-2fbdd2a87dba-kube-api-access-64ssc\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.540234 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f242653b-4449-452c-a438-2fbdd2a87dba-combined-ca-bundle\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.540265 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f242653b-4449-452c-a438-2fbdd2a87dba-httpd-config\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.540300 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f242653b-4449-452c-a438-2fbdd2a87dba-config\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.545885 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f242653b-4449-452c-a438-2fbdd2a87dba-httpd-config\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.545910 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f242653b-4449-452c-a438-2fbdd2a87dba-combined-ca-bundle\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.546458 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f242653b-4449-452c-a438-2fbdd2a87dba-config\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.564283 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64ssc\" (UniqueName: \"kubernetes.io/projected/f242653b-4449-452c-a438-2fbdd2a87dba-kube-api-access-64ssc\") pod \"neutron-65f4496dcc-6mkfj\" (UID: \"f242653b-4449-452c-a438-2fbdd2a87dba\") " pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.650516 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:31 crc kubenswrapper[4871]: I1007 23:40:31.961147 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b7d74c6cf-5w7gm"] Oct 07 23:40:32 crc kubenswrapper[4871]: I1007 23:40:32.025884 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" event={"ID":"e228a39c-1929-416e-a1ab-15c66c6eb329","Type":"ContainerStarted","Data":"34bfb90b579acf96b0aabb3774119f95fad3689e68aa7a12bcf494c5c09b54e9"} Oct 07 23:40:32 crc kubenswrapper[4871]: I1007 23:40:32.268445 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65f4496dcc-6mkfj"] Oct 07 23:40:32 crc kubenswrapper[4871]: W1007 23:40:32.271998 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf242653b_4449_452c_a438_2fbdd2a87dba.slice/crio-10f1ac0fee53ae597597408f7e0f3863171469fd159dbbc13a21d60b337f6842 WatchSource:0}: Error finding container 10f1ac0fee53ae597597408f7e0f3863171469fd159dbbc13a21d60b337f6842: Status 404 returned error can't find the container with id 10f1ac0fee53ae597597408f7e0f3863171469fd159dbbc13a21d60b337f6842 Oct 07 23:40:33 crc kubenswrapper[4871]: I1007 23:40:33.037075 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65f4496dcc-6mkfj" event={"ID":"f242653b-4449-452c-a438-2fbdd2a87dba","Type":"ContainerStarted","Data":"3cee56cfdc0c0c3f441183c0d7de48e6f0d3a413163fefa0c46471d5148a25bc"} Oct 07 23:40:33 crc kubenswrapper[4871]: I1007 23:40:33.037349 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65f4496dcc-6mkfj" event={"ID":"f242653b-4449-452c-a438-2fbdd2a87dba","Type":"ContainerStarted","Data":"375ab4ff1b713603014b20f3065638e3cac82cbd5e837008d77f113d724e08df"} Oct 07 23:40:33 crc kubenswrapper[4871]: I1007 23:40:33.037359 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65f4496dcc-6mkfj" event={"ID":"f242653b-4449-452c-a438-2fbdd2a87dba","Type":"ContainerStarted","Data":"10f1ac0fee53ae597597408f7e0f3863171469fd159dbbc13a21d60b337f6842"} Oct 07 23:40:33 crc kubenswrapper[4871]: I1007 23:40:33.037372 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:40:33 crc kubenswrapper[4871]: I1007 23:40:33.039996 4871 generic.go:334] "Generic (PLEG): container finished" podID="e228a39c-1929-416e-a1ab-15c66c6eb329" containerID="656bb477211bd3649b1335ee2e4242b1ec560e2171baf90587abca189766b510" exitCode=0 Oct 07 23:40:33 crc kubenswrapper[4871]: I1007 23:40:33.040051 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" event={"ID":"e228a39c-1929-416e-a1ab-15c66c6eb329","Type":"ContainerDied","Data":"656bb477211bd3649b1335ee2e4242b1ec560e2171baf90587abca189766b510"} Oct 07 23:40:33 crc kubenswrapper[4871]: I1007 23:40:33.068145 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-65f4496dcc-6mkfj" podStartSLOduration=2.068123843 podStartE2EDuration="2.068123843s" podCreationTimestamp="2025-10-07 23:40:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:40:33.055955932 +0000 UTC m=+5506.858654035" watchObservedRunningTime="2025-10-07 23:40:33.068123843 +0000 UTC m=+5506.870821906" Oct 07 23:40:34 crc kubenswrapper[4871]: I1007 23:40:34.058446 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" event={"ID":"e228a39c-1929-416e-a1ab-15c66c6eb329","Type":"ContainerStarted","Data":"0e19ce0ccba6c04e71df39ba8e9b1de79f106ec501428817450cb55b52a9bd24"} Oct 07 23:40:34 crc kubenswrapper[4871]: I1007 23:40:34.088050 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" podStartSLOduration=3.088025608 podStartE2EDuration="3.088025608s" podCreationTimestamp="2025-10-07 23:40:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:40:34.082540653 +0000 UTC m=+5507.885238756" watchObservedRunningTime="2025-10-07 23:40:34.088025608 +0000 UTC m=+5507.890723721" Oct 07 23:40:35 crc kubenswrapper[4871]: I1007 23:40:35.068746 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:35 crc kubenswrapper[4871]: I1007 23:40:35.512785 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:40:35 crc kubenswrapper[4871]: I1007 23:40:35.512925 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:40:35 crc kubenswrapper[4871]: I1007 23:40:35.512989 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:40:35 crc kubenswrapper[4871]: I1007 23:40:35.514053 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a7d7ec4f805a57b5cbdcea4bbeab976af5520b960d9f0ff723e036600ebfeb03"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:40:35 crc kubenswrapper[4871]: I1007 23:40:35.514200 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://a7d7ec4f805a57b5cbdcea4bbeab976af5520b960d9f0ff723e036600ebfeb03" gracePeriod=600 Oct 07 23:40:36 crc kubenswrapper[4871]: I1007 23:40:36.078778 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="a7d7ec4f805a57b5cbdcea4bbeab976af5520b960d9f0ff723e036600ebfeb03" exitCode=0 Oct 07 23:40:36 crc kubenswrapper[4871]: I1007 23:40:36.078865 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"a7d7ec4f805a57b5cbdcea4bbeab976af5520b960d9f0ff723e036600ebfeb03"} Oct 07 23:40:36 crc kubenswrapper[4871]: I1007 23:40:36.079098 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5"} Oct 07 23:40:36 crc kubenswrapper[4871]: I1007 23:40:36.079124 4871 scope.go:117] "RemoveContainer" containerID="f773cd42a6efc52e95d3ef7f1785b8d4c7f9e42feb39b8df9dbbb4211e76fc9a" Oct 07 23:40:41 crc kubenswrapper[4871]: I1007 23:40:41.509088 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:40:41 crc kubenswrapper[4871]: I1007 23:40:41.606771 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d44dfd697-wnr8b"] Oct 07 23:40:41 crc kubenswrapper[4871]: I1007 23:40:41.607125 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" podUID="5548ed74-08b9-42c3-8d2c-99e82f860a92" containerName="dnsmasq-dns" containerID="cri-o://2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f" gracePeriod=10 Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.055622 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.144058 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5zpz\" (UniqueName: \"kubernetes.io/projected/5548ed74-08b9-42c3-8d2c-99e82f860a92-kube-api-access-q5zpz\") pod \"5548ed74-08b9-42c3-8d2c-99e82f860a92\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.144134 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-sb\") pod \"5548ed74-08b9-42c3-8d2c-99e82f860a92\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.144227 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-nb\") pod \"5548ed74-08b9-42c3-8d2c-99e82f860a92\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.144264 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-config\") pod \"5548ed74-08b9-42c3-8d2c-99e82f860a92\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.144388 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-dns-svc\") pod \"5548ed74-08b9-42c3-8d2c-99e82f860a92\" (UID: \"5548ed74-08b9-42c3-8d2c-99e82f860a92\") " Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.148613 4871 generic.go:334] "Generic (PLEG): container finished" podID="5548ed74-08b9-42c3-8d2c-99e82f860a92" containerID="2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f" exitCode=0 Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.148753 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" event={"ID":"5548ed74-08b9-42c3-8d2c-99e82f860a92","Type":"ContainerDied","Data":"2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f"} Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.148779 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" event={"ID":"5548ed74-08b9-42c3-8d2c-99e82f860a92","Type":"ContainerDied","Data":"9988ce2323a94fa1876ce8febd438555517d419fec5bd7d09c73e50a16345974"} Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.148809 4871 scope.go:117] "RemoveContainer" containerID="2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.148936 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d44dfd697-wnr8b" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.154146 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5548ed74-08b9-42c3-8d2c-99e82f860a92-kube-api-access-q5zpz" (OuterVolumeSpecName: "kube-api-access-q5zpz") pod "5548ed74-08b9-42c3-8d2c-99e82f860a92" (UID: "5548ed74-08b9-42c3-8d2c-99e82f860a92"). InnerVolumeSpecName "kube-api-access-q5zpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.189543 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5548ed74-08b9-42c3-8d2c-99e82f860a92" (UID: "5548ed74-08b9-42c3-8d2c-99e82f860a92"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.201144 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-config" (OuterVolumeSpecName: "config") pod "5548ed74-08b9-42c3-8d2c-99e82f860a92" (UID: "5548ed74-08b9-42c3-8d2c-99e82f860a92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.206307 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5548ed74-08b9-42c3-8d2c-99e82f860a92" (UID: "5548ed74-08b9-42c3-8d2c-99e82f860a92"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.210742 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5548ed74-08b9-42c3-8d2c-99e82f860a92" (UID: "5548ed74-08b9-42c3-8d2c-99e82f860a92"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.246539 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.246579 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5zpz\" (UniqueName: \"kubernetes.io/projected/5548ed74-08b9-42c3-8d2c-99e82f860a92-kube-api-access-q5zpz\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.246597 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.246609 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.246622 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548ed74-08b9-42c3-8d2c-99e82f860a92-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.258620 4871 scope.go:117] "RemoveContainer" containerID="f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.294392 4871 scope.go:117] "RemoveContainer" containerID="2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f" Oct 07 23:40:42 crc kubenswrapper[4871]: E1007 23:40:42.294820 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f\": container with ID starting with 2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f not found: ID does not exist" containerID="2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.294862 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f"} err="failed to get container status \"2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f\": rpc error: code = NotFound desc = could not find container \"2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f\": container with ID starting with 2001ddd447f195e8bd15ab100a4f90b8175e5ff765b38e1005564384857e794f not found: ID does not exist" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.294888 4871 scope.go:117] "RemoveContainer" containerID="f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503" Oct 07 23:40:42 crc kubenswrapper[4871]: E1007 23:40:42.295244 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503\": container with ID starting with f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503 not found: ID does not exist" containerID="f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.295297 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503"} err="failed to get container status \"f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503\": rpc error: code = NotFound desc = could not find container \"f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503\": container with ID starting with f82b2efe777bf7039a6a63cddf8561a7b5129db5238c3332421df8158d9f1503 not found: ID does not exist" Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.485042 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d44dfd697-wnr8b"] Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.497348 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d44dfd697-wnr8b"] Oct 07 23:40:42 crc kubenswrapper[4871]: I1007 23:40:42.998351 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5548ed74-08b9-42c3-8d2c-99e82f860a92" path="/var/lib/kubelet/pods/5548ed74-08b9-42c3-8d2c-99e82f860a92/volumes" Oct 07 23:41:01 crc kubenswrapper[4871]: I1007 23:41:01.668006 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-65f4496dcc-6mkfj" Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.351302 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-74vwd"] Oct 07 23:41:12 crc kubenswrapper[4871]: E1007 23:41:12.352576 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5548ed74-08b9-42c3-8d2c-99e82f860a92" containerName="dnsmasq-dns" Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.352597 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5548ed74-08b9-42c3-8d2c-99e82f860a92" containerName="dnsmasq-dns" Oct 07 23:41:12 crc kubenswrapper[4871]: E1007 23:41:12.352626 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5548ed74-08b9-42c3-8d2c-99e82f860a92" containerName="init" Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.352637 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5548ed74-08b9-42c3-8d2c-99e82f860a92" containerName="init" Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.352953 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5548ed74-08b9-42c3-8d2c-99e82f860a92" containerName="dnsmasq-dns" Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.353901 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74vwd" Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.366657 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-74vwd"] Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.416890 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfwhx\" (UniqueName: \"kubernetes.io/projected/ed609edd-a961-4533-a12d-11adc1800513-kube-api-access-pfwhx\") pod \"glance-db-create-74vwd\" (UID: \"ed609edd-a961-4533-a12d-11adc1800513\") " pod="openstack/glance-db-create-74vwd" Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.518905 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfwhx\" (UniqueName: \"kubernetes.io/projected/ed609edd-a961-4533-a12d-11adc1800513-kube-api-access-pfwhx\") pod \"glance-db-create-74vwd\" (UID: \"ed609edd-a961-4533-a12d-11adc1800513\") " pod="openstack/glance-db-create-74vwd" Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.539495 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfwhx\" (UniqueName: \"kubernetes.io/projected/ed609edd-a961-4533-a12d-11adc1800513-kube-api-access-pfwhx\") pod \"glance-db-create-74vwd\" (UID: \"ed609edd-a961-4533-a12d-11adc1800513\") " pod="openstack/glance-db-create-74vwd" Oct 07 23:41:12 crc kubenswrapper[4871]: I1007 23:41:12.682780 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74vwd" Oct 07 23:41:13 crc kubenswrapper[4871]: I1007 23:41:13.121583 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-74vwd"] Oct 07 23:41:13 crc kubenswrapper[4871]: I1007 23:41:13.451924 4871 generic.go:334] "Generic (PLEG): container finished" podID="ed609edd-a961-4533-a12d-11adc1800513" containerID="67eaf865e8d6d15aacd921116b15cab4bc1d0e2fe81329a467310321e0c57244" exitCode=0 Oct 07 23:41:13 crc kubenswrapper[4871]: I1007 23:41:13.452029 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74vwd" event={"ID":"ed609edd-a961-4533-a12d-11adc1800513","Type":"ContainerDied","Data":"67eaf865e8d6d15aacd921116b15cab4bc1d0e2fe81329a467310321e0c57244"} Oct 07 23:41:13 crc kubenswrapper[4871]: I1007 23:41:13.452303 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74vwd" event={"ID":"ed609edd-a961-4533-a12d-11adc1800513","Type":"ContainerStarted","Data":"d61a29030b59a9e2170cbe06f438b9191bdc0e4b4fb20f191cb504b19ba2e334"} Oct 07 23:41:14 crc kubenswrapper[4871]: I1007 23:41:14.735277 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74vwd" Oct 07 23:41:14 crc kubenswrapper[4871]: I1007 23:41:14.857433 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfwhx\" (UniqueName: \"kubernetes.io/projected/ed609edd-a961-4533-a12d-11adc1800513-kube-api-access-pfwhx\") pod \"ed609edd-a961-4533-a12d-11adc1800513\" (UID: \"ed609edd-a961-4533-a12d-11adc1800513\") " Oct 07 23:41:14 crc kubenswrapper[4871]: I1007 23:41:14.871154 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed609edd-a961-4533-a12d-11adc1800513-kube-api-access-pfwhx" (OuterVolumeSpecName: "kube-api-access-pfwhx") pod "ed609edd-a961-4533-a12d-11adc1800513" (UID: "ed609edd-a961-4533-a12d-11adc1800513"). InnerVolumeSpecName "kube-api-access-pfwhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:41:14 crc kubenswrapper[4871]: I1007 23:41:14.959543 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfwhx\" (UniqueName: \"kubernetes.io/projected/ed609edd-a961-4533-a12d-11adc1800513-kube-api-access-pfwhx\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:15 crc kubenswrapper[4871]: I1007 23:41:15.468761 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74vwd" event={"ID":"ed609edd-a961-4533-a12d-11adc1800513","Type":"ContainerDied","Data":"d61a29030b59a9e2170cbe06f438b9191bdc0e4b4fb20f191cb504b19ba2e334"} Oct 07 23:41:15 crc kubenswrapper[4871]: I1007 23:41:15.468844 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d61a29030b59a9e2170cbe06f438b9191bdc0e4b4fb20f191cb504b19ba2e334" Oct 07 23:41:15 crc kubenswrapper[4871]: I1007 23:41:15.468869 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74vwd" Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.493282 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-557c-account-create-8mc29"] Oct 07 23:41:22 crc kubenswrapper[4871]: E1007 23:41:22.494259 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed609edd-a961-4533-a12d-11adc1800513" containerName="mariadb-database-create" Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.494276 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed609edd-a961-4533-a12d-11adc1800513" containerName="mariadb-database-create" Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.494468 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed609edd-a961-4533-a12d-11adc1800513" containerName="mariadb-database-create" Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.495163 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-557c-account-create-8mc29" Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.504589 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.529961 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-557c-account-create-8mc29"] Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.607294 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z6rg\" (UniqueName: \"kubernetes.io/projected/cbc912d0-51de-4f35-85b3-2f3403d2c554-kube-api-access-5z6rg\") pod \"glance-557c-account-create-8mc29\" (UID: \"cbc912d0-51de-4f35-85b3-2f3403d2c554\") " pod="openstack/glance-557c-account-create-8mc29" Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.709835 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z6rg\" (UniqueName: \"kubernetes.io/projected/cbc912d0-51de-4f35-85b3-2f3403d2c554-kube-api-access-5z6rg\") pod \"glance-557c-account-create-8mc29\" (UID: \"cbc912d0-51de-4f35-85b3-2f3403d2c554\") " pod="openstack/glance-557c-account-create-8mc29" Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.732576 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z6rg\" (UniqueName: \"kubernetes.io/projected/cbc912d0-51de-4f35-85b3-2f3403d2c554-kube-api-access-5z6rg\") pod \"glance-557c-account-create-8mc29\" (UID: \"cbc912d0-51de-4f35-85b3-2f3403d2c554\") " pod="openstack/glance-557c-account-create-8mc29" Oct 07 23:41:22 crc kubenswrapper[4871]: I1007 23:41:22.819523 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-557c-account-create-8mc29" Oct 07 23:41:23 crc kubenswrapper[4871]: I1007 23:41:23.318941 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-557c-account-create-8mc29"] Oct 07 23:41:23 crc kubenswrapper[4871]: W1007 23:41:23.323415 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbc912d0_51de_4f35_85b3_2f3403d2c554.slice/crio-a3c73d65e4efdfd853126c350f294259322be074404b3368323d330eae7f0c2e WatchSource:0}: Error finding container a3c73d65e4efdfd853126c350f294259322be074404b3368323d330eae7f0c2e: Status 404 returned error can't find the container with id a3c73d65e4efdfd853126c350f294259322be074404b3368323d330eae7f0c2e Oct 07 23:41:23 crc kubenswrapper[4871]: I1007 23:41:23.551264 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-557c-account-create-8mc29" event={"ID":"cbc912d0-51de-4f35-85b3-2f3403d2c554","Type":"ContainerStarted","Data":"7453dba567a5dcc87ebe1c6e53a79f688d273ce2f7a2981e5d483e37fe9027b3"} Oct 07 23:41:23 crc kubenswrapper[4871]: I1007 23:41:23.551508 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-557c-account-create-8mc29" event={"ID":"cbc912d0-51de-4f35-85b3-2f3403d2c554","Type":"ContainerStarted","Data":"a3c73d65e4efdfd853126c350f294259322be074404b3368323d330eae7f0c2e"} Oct 07 23:41:23 crc kubenswrapper[4871]: I1007 23:41:23.579607 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-557c-account-create-8mc29" podStartSLOduration=1.579577338 podStartE2EDuration="1.579577338s" podCreationTimestamp="2025-10-07 23:41:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:41:23.570417826 +0000 UTC m=+5557.373115959" watchObservedRunningTime="2025-10-07 23:41:23.579577338 +0000 UTC m=+5557.382275451" Oct 07 23:41:24 crc kubenswrapper[4871]: I1007 23:41:24.566226 4871 generic.go:334] "Generic (PLEG): container finished" podID="cbc912d0-51de-4f35-85b3-2f3403d2c554" containerID="7453dba567a5dcc87ebe1c6e53a79f688d273ce2f7a2981e5d483e37fe9027b3" exitCode=0 Oct 07 23:41:24 crc kubenswrapper[4871]: I1007 23:41:24.566301 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-557c-account-create-8mc29" event={"ID":"cbc912d0-51de-4f35-85b3-2f3403d2c554","Type":"ContainerDied","Data":"7453dba567a5dcc87ebe1c6e53a79f688d273ce2f7a2981e5d483e37fe9027b3"} Oct 07 23:41:26 crc kubenswrapper[4871]: I1007 23:41:26.018481 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-557c-account-create-8mc29" Oct 07 23:41:26 crc kubenswrapper[4871]: I1007 23:41:26.072224 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z6rg\" (UniqueName: \"kubernetes.io/projected/cbc912d0-51de-4f35-85b3-2f3403d2c554-kube-api-access-5z6rg\") pod \"cbc912d0-51de-4f35-85b3-2f3403d2c554\" (UID: \"cbc912d0-51de-4f35-85b3-2f3403d2c554\") " Oct 07 23:41:26 crc kubenswrapper[4871]: I1007 23:41:26.080761 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbc912d0-51de-4f35-85b3-2f3403d2c554-kube-api-access-5z6rg" (OuterVolumeSpecName: "kube-api-access-5z6rg") pod "cbc912d0-51de-4f35-85b3-2f3403d2c554" (UID: "cbc912d0-51de-4f35-85b3-2f3403d2c554"). InnerVolumeSpecName "kube-api-access-5z6rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:41:26 crc kubenswrapper[4871]: I1007 23:41:26.173948 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z6rg\" (UniqueName: \"kubernetes.io/projected/cbc912d0-51de-4f35-85b3-2f3403d2c554-kube-api-access-5z6rg\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:26 crc kubenswrapper[4871]: I1007 23:41:26.605589 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-557c-account-create-8mc29" event={"ID":"cbc912d0-51de-4f35-85b3-2f3403d2c554","Type":"ContainerDied","Data":"a3c73d65e4efdfd853126c350f294259322be074404b3368323d330eae7f0c2e"} Oct 07 23:41:26 crc kubenswrapper[4871]: I1007 23:41:26.605653 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3c73d65e4efdfd853126c350f294259322be074404b3368323d330eae7f0c2e" Oct 07 23:41:26 crc kubenswrapper[4871]: I1007 23:41:26.605752 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-557c-account-create-8mc29" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.733886 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-7vdvf"] Oct 07 23:41:27 crc kubenswrapper[4871]: E1007 23:41:27.734297 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc912d0-51de-4f35-85b3-2f3403d2c554" containerName="mariadb-account-create" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.734309 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc912d0-51de-4f35-85b3-2f3403d2c554" containerName="mariadb-account-create" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.734491 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc912d0-51de-4f35-85b3-2f3403d2c554" containerName="mariadb-account-create" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.735143 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.738880 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.742735 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-7vdvf"] Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.746551 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-m6mm9" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.804647 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-config-data\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.805112 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-combined-ca-bundle\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.805172 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rszpl\" (UniqueName: \"kubernetes.io/projected/f4b06d89-dec9-4efd-9e84-be7b357e916a-kube-api-access-rszpl\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.805235 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-db-sync-config-data\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.905923 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-db-sync-config-data\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.905969 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-config-data\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.906056 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-combined-ca-bundle\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.906097 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rszpl\" (UniqueName: \"kubernetes.io/projected/f4b06d89-dec9-4efd-9e84-be7b357e916a-kube-api-access-rszpl\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.912623 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-db-sync-config-data\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.914884 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-config-data\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.921665 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-combined-ca-bundle\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:27 crc kubenswrapper[4871]: I1007 23:41:27.923762 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rszpl\" (UniqueName: \"kubernetes.io/projected/f4b06d89-dec9-4efd-9e84-be7b357e916a-kube-api-access-rszpl\") pod \"glance-db-sync-7vdvf\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:28 crc kubenswrapper[4871]: I1007 23:41:28.068161 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:28 crc kubenswrapper[4871]: I1007 23:41:28.554615 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-7vdvf"] Oct 07 23:41:28 crc kubenswrapper[4871]: I1007 23:41:28.627049 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7vdvf" event={"ID":"f4b06d89-dec9-4efd-9e84-be7b357e916a","Type":"ContainerStarted","Data":"5b1eb122b1458bf61b288568c1029072bea7c4553599ec04204025a4d337e987"} Oct 07 23:41:29 crc kubenswrapper[4871]: I1007 23:41:29.639865 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7vdvf" event={"ID":"f4b06d89-dec9-4efd-9e84-be7b357e916a","Type":"ContainerStarted","Data":"f9baa490831c655023bad085dd44139134fa574578421a343dbdf99605b825d2"} Oct 07 23:41:29 crc kubenswrapper[4871]: I1007 23:41:29.663416 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-7vdvf" podStartSLOduration=2.663398815 podStartE2EDuration="2.663398815s" podCreationTimestamp="2025-10-07 23:41:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:41:29.657025857 +0000 UTC m=+5563.459723930" watchObservedRunningTime="2025-10-07 23:41:29.663398815 +0000 UTC m=+5563.466096888" Oct 07 23:41:32 crc kubenswrapper[4871]: I1007 23:41:32.671157 4871 generic.go:334] "Generic (PLEG): container finished" podID="f4b06d89-dec9-4efd-9e84-be7b357e916a" containerID="f9baa490831c655023bad085dd44139134fa574578421a343dbdf99605b825d2" exitCode=0 Oct 07 23:41:32 crc kubenswrapper[4871]: I1007 23:41:32.671263 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7vdvf" event={"ID":"f4b06d89-dec9-4efd-9e84-be7b357e916a","Type":"ContainerDied","Data":"f9baa490831c655023bad085dd44139134fa574578421a343dbdf99605b825d2"} Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.152714 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.240399 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rszpl\" (UniqueName: \"kubernetes.io/projected/f4b06d89-dec9-4efd-9e84-be7b357e916a-kube-api-access-rszpl\") pod \"f4b06d89-dec9-4efd-9e84-be7b357e916a\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.240666 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-combined-ca-bundle\") pod \"f4b06d89-dec9-4efd-9e84-be7b357e916a\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.240694 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-db-sync-config-data\") pod \"f4b06d89-dec9-4efd-9e84-be7b357e916a\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.241587 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-config-data\") pod \"f4b06d89-dec9-4efd-9e84-be7b357e916a\" (UID: \"f4b06d89-dec9-4efd-9e84-be7b357e916a\") " Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.249962 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f4b06d89-dec9-4efd-9e84-be7b357e916a" (UID: "f4b06d89-dec9-4efd-9e84-be7b357e916a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.252667 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4b06d89-dec9-4efd-9e84-be7b357e916a-kube-api-access-rszpl" (OuterVolumeSpecName: "kube-api-access-rszpl") pod "f4b06d89-dec9-4efd-9e84-be7b357e916a" (UID: "f4b06d89-dec9-4efd-9e84-be7b357e916a"). InnerVolumeSpecName "kube-api-access-rszpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.281677 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4b06d89-dec9-4efd-9e84-be7b357e916a" (UID: "f4b06d89-dec9-4efd-9e84-be7b357e916a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.309324 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-config-data" (OuterVolumeSpecName: "config-data") pod "f4b06d89-dec9-4efd-9e84-be7b357e916a" (UID: "f4b06d89-dec9-4efd-9e84-be7b357e916a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.344118 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.344166 4871 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.344184 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b06d89-dec9-4efd-9e84-be7b357e916a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.344204 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rszpl\" (UniqueName: \"kubernetes.io/projected/f4b06d89-dec9-4efd-9e84-be7b357e916a-kube-api-access-rszpl\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.698232 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7vdvf" event={"ID":"f4b06d89-dec9-4efd-9e84-be7b357e916a","Type":"ContainerDied","Data":"5b1eb122b1458bf61b288568c1029072bea7c4553599ec04204025a4d337e987"} Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.698597 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b1eb122b1458bf61b288568c1029072bea7c4553599ec04204025a4d337e987" Oct 07 23:41:34 crc kubenswrapper[4871]: I1007 23:41:34.698300 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7vdvf" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.161975 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bcbfc77f9-sljjb"] Oct 07 23:41:35 crc kubenswrapper[4871]: E1007 23:41:35.162563 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b06d89-dec9-4efd-9e84-be7b357e916a" containerName="glance-db-sync" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.162582 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b06d89-dec9-4efd-9e84-be7b357e916a" containerName="glance-db-sync" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.162769 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b06d89-dec9-4efd-9e84-be7b357e916a" containerName="glance-db-sync" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.163718 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.185535 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bcbfc77f9-sljjb"] Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.195769 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.200607 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.206111 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.206439 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-m6mm9" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.206492 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.207479 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.230593 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.268910 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-nb\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269003 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-scripts\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269023 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shvmr\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-kube-api-access-shvmr\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269062 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269091 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-logs\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269112 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-ceph\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269126 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-sb\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269148 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgsxk\" (UniqueName: \"kubernetes.io/projected/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-kube-api-access-bgsxk\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269170 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-config\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269196 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269214 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-dns-svc\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.269233 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-config-data\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.286849 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.289558 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.292400 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.297215 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.370927 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.370972 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-dns-svc\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371002 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-config-data\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371038 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-logs\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371071 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-nb\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371089 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371114 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371150 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371173 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371201 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371224 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-scripts\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371255 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shvmr\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-kube-api-access-shvmr\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371286 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shh5g\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-kube-api-access-shh5g\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371313 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371347 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-logs\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371365 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-ceph\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371380 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-sb\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371398 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgsxk\" (UniqueName: \"kubernetes.io/projected/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-kube-api-access-bgsxk\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.371419 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-config\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.372329 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-config\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.372342 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.372537 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-logs\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.373188 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-sb\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.373188 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-dns-svc\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.373297 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-nb\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.376645 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-ceph\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.381827 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-scripts\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.383894 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-config-data\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.394520 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shvmr\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-kube-api-access-shvmr\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.395387 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.396217 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgsxk\" (UniqueName: \"kubernetes.io/projected/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-kube-api-access-bgsxk\") pod \"dnsmasq-dns-6bcbfc77f9-sljjb\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.473355 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-logs\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.473411 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.473442 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.473475 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.473499 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.473527 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.473562 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shh5g\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-kube-api-access-shh5g\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.475006 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-logs\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.475340 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.480094 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.485361 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.486605 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.492415 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.495504 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.496164 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shh5g\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-kube-api-access-shh5g\") pod \"glance-default-internal-api-0\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.528339 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.605439 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:35 crc kubenswrapper[4871]: I1007 23:41:35.951000 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bcbfc77f9-sljjb"] Oct 07 23:41:36 crc kubenswrapper[4871]: I1007 23:41:36.180837 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:41:36 crc kubenswrapper[4871]: W1007 23:41:36.183553 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14a0eb07_e443_4829_9d12_4832818ccaea.slice/crio-ba024a504c9afd9c73aae74308c4a6f5ba34fa0e4ef64b5da8c1d29644898517 WatchSource:0}: Error finding container ba024a504c9afd9c73aae74308c4a6f5ba34fa0e4ef64b5da8c1d29644898517: Status 404 returned error can't find the container with id ba024a504c9afd9c73aae74308c4a6f5ba34fa0e4ef64b5da8c1d29644898517 Oct 07 23:41:36 crc kubenswrapper[4871]: I1007 23:41:36.313837 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:41:36 crc kubenswrapper[4871]: I1007 23:41:36.516972 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:41:36 crc kubenswrapper[4871]: I1007 23:41:36.724426 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f100e6ff-78fe-4c67-b7d8-050e4a07ad40","Type":"ContainerStarted","Data":"78f937ada2c2ab0c715d24c503751db691afc685f80827731ace562a15aff75d"} Oct 07 23:41:36 crc kubenswrapper[4871]: I1007 23:41:36.725988 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14a0eb07-e443-4829-9d12-4832818ccaea","Type":"ContainerStarted","Data":"644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76"} Oct 07 23:41:36 crc kubenswrapper[4871]: I1007 23:41:36.726013 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14a0eb07-e443-4829-9d12-4832818ccaea","Type":"ContainerStarted","Data":"ba024a504c9afd9c73aae74308c4a6f5ba34fa0e4ef64b5da8c1d29644898517"} Oct 07 23:41:36 crc kubenswrapper[4871]: I1007 23:41:36.727601 4871 generic.go:334] "Generic (PLEG): container finished" podID="336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" containerID="8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf" exitCode=0 Oct 07 23:41:36 crc kubenswrapper[4871]: I1007 23:41:36.727645 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" event={"ID":"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3","Type":"ContainerDied","Data":"8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf"} Oct 07 23:41:36 crc kubenswrapper[4871]: I1007 23:41:36.727670 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" event={"ID":"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3","Type":"ContainerStarted","Data":"c3371ab2a17405621a065f1f2ba27a462730a494d5647f27b77ca405810d75b0"} Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.737129 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" event={"ID":"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3","Type":"ContainerStarted","Data":"16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad"} Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.737774 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.739157 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f100e6ff-78fe-4c67-b7d8-050e4a07ad40","Type":"ContainerStarted","Data":"af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1"} Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.739188 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f100e6ff-78fe-4c67-b7d8-050e4a07ad40","Type":"ContainerStarted","Data":"232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447"} Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.742089 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14a0eb07-e443-4829-9d12-4832818ccaea","Type":"ContainerStarted","Data":"bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216"} Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.742223 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="14a0eb07-e443-4829-9d12-4832818ccaea" containerName="glance-log" containerID="cri-o://644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76" gracePeriod=30 Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.743898 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="14a0eb07-e443-4829-9d12-4832818ccaea" containerName="glance-httpd" containerID="cri-o://bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216" gracePeriod=30 Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.763363 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" podStartSLOduration=2.763342218 podStartE2EDuration="2.763342218s" podCreationTimestamp="2025-10-07 23:41:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:41:37.756826196 +0000 UTC m=+5571.559524279" watchObservedRunningTime="2025-10-07 23:41:37.763342218 +0000 UTC m=+5571.566040291" Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.776931 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.776910476 podStartE2EDuration="2.776910476s" podCreationTimestamp="2025-10-07 23:41:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:41:37.776519756 +0000 UTC m=+5571.579217829" watchObservedRunningTime="2025-10-07 23:41:37.776910476 +0000 UTC m=+5571.579608559" Oct 07 23:41:37 crc kubenswrapper[4871]: I1007 23:41:37.814095 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.814072537 podStartE2EDuration="2.814072537s" podCreationTimestamp="2025-10-07 23:41:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:41:37.805411028 +0000 UTC m=+5571.608109111" watchObservedRunningTime="2025-10-07 23:41:37.814072537 +0000 UTC m=+5571.616770610" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.402457 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.444393 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-logs\") pod \"14a0eb07-e443-4829-9d12-4832818ccaea\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.444439 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-config-data\") pod \"14a0eb07-e443-4829-9d12-4832818ccaea\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.444524 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-ceph\") pod \"14a0eb07-e443-4829-9d12-4832818ccaea\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.444545 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-scripts\") pod \"14a0eb07-e443-4829-9d12-4832818ccaea\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.444599 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-combined-ca-bundle\") pod \"14a0eb07-e443-4829-9d12-4832818ccaea\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.444697 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-httpd-run\") pod \"14a0eb07-e443-4829-9d12-4832818ccaea\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.444720 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shvmr\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-kube-api-access-shvmr\") pod \"14a0eb07-e443-4829-9d12-4832818ccaea\" (UID: \"14a0eb07-e443-4829-9d12-4832818ccaea\") " Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.445035 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-logs" (OuterVolumeSpecName: "logs") pod "14a0eb07-e443-4829-9d12-4832818ccaea" (UID: "14a0eb07-e443-4829-9d12-4832818ccaea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.446471 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "14a0eb07-e443-4829-9d12-4832818ccaea" (UID: "14a0eb07-e443-4829-9d12-4832818ccaea"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.451015 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-kube-api-access-shvmr" (OuterVolumeSpecName: "kube-api-access-shvmr") pod "14a0eb07-e443-4829-9d12-4832818ccaea" (UID: "14a0eb07-e443-4829-9d12-4832818ccaea"). InnerVolumeSpecName "kube-api-access-shvmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.452601 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-ceph" (OuterVolumeSpecName: "ceph") pod "14a0eb07-e443-4829-9d12-4832818ccaea" (UID: "14a0eb07-e443-4829-9d12-4832818ccaea"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.453321 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-scripts" (OuterVolumeSpecName: "scripts") pod "14a0eb07-e443-4829-9d12-4832818ccaea" (UID: "14a0eb07-e443-4829-9d12-4832818ccaea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.481542 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14a0eb07-e443-4829-9d12-4832818ccaea" (UID: "14a0eb07-e443-4829-9d12-4832818ccaea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.518299 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-config-data" (OuterVolumeSpecName: "config-data") pod "14a0eb07-e443-4829-9d12-4832818ccaea" (UID: "14a0eb07-e443-4829-9d12-4832818ccaea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.546286 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.546319 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shvmr\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-kube-api-access-shvmr\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.546330 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a0eb07-e443-4829-9d12-4832818ccaea-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.546338 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.546346 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/14a0eb07-e443-4829-9d12-4832818ccaea-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.546353 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.546361 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a0eb07-e443-4829-9d12-4832818ccaea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.761526 4871 generic.go:334] "Generic (PLEG): container finished" podID="14a0eb07-e443-4829-9d12-4832818ccaea" containerID="bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216" exitCode=0 Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.761554 4871 generic.go:334] "Generic (PLEG): container finished" podID="14a0eb07-e443-4829-9d12-4832818ccaea" containerID="644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76" exitCode=143 Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.762556 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.763986 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14a0eb07-e443-4829-9d12-4832818ccaea","Type":"ContainerDied","Data":"bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216"} Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.764038 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14a0eb07-e443-4829-9d12-4832818ccaea","Type":"ContainerDied","Data":"644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76"} Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.764056 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"14a0eb07-e443-4829-9d12-4832818ccaea","Type":"ContainerDied","Data":"ba024a504c9afd9c73aae74308c4a6f5ba34fa0e4ef64b5da8c1d29644898517"} Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.764076 4871 scope.go:117] "RemoveContainer" containerID="bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.790473 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.792605 4871 scope.go:117] "RemoveContainer" containerID="644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.802228 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.810232 4871 scope.go:117] "RemoveContainer" containerID="bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216" Oct 07 23:41:38 crc kubenswrapper[4871]: E1007 23:41:38.810674 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216\": container with ID starting with bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216 not found: ID does not exist" containerID="bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.810715 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216"} err="failed to get container status \"bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216\": rpc error: code = NotFound desc = could not find container \"bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216\": container with ID starting with bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216 not found: ID does not exist" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.810742 4871 scope.go:117] "RemoveContainer" containerID="644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76" Oct 07 23:41:38 crc kubenswrapper[4871]: E1007 23:41:38.811024 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76\": container with ID starting with 644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76 not found: ID does not exist" containerID="644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.811054 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76"} err="failed to get container status \"644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76\": rpc error: code = NotFound desc = could not find container \"644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76\": container with ID starting with 644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76 not found: ID does not exist" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.811072 4871 scope.go:117] "RemoveContainer" containerID="bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.811341 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216"} err="failed to get container status \"bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216\": rpc error: code = NotFound desc = could not find container \"bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216\": container with ID starting with bb2e03c3fa80b6e40d16aa76a896c55bfc094105bef045bf452f4463d48d8216 not found: ID does not exist" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.811688 4871 scope.go:117] "RemoveContainer" containerID="644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.811935 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76"} err="failed to get container status \"644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76\": rpc error: code = NotFound desc = could not find container \"644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76\": container with ID starting with 644575dd4999d8ac030d08fa3d920a6fc97110b15a8b6deb337b3419efc1ef76 not found: ID does not exist" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.826648 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:41:38 crc kubenswrapper[4871]: E1007 23:41:38.827115 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a0eb07-e443-4829-9d12-4832818ccaea" containerName="glance-log" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.827139 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a0eb07-e443-4829-9d12-4832818ccaea" containerName="glance-log" Oct 07 23:41:38 crc kubenswrapper[4871]: E1007 23:41:38.827189 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a0eb07-e443-4829-9d12-4832818ccaea" containerName="glance-httpd" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.827198 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a0eb07-e443-4829-9d12-4832818ccaea" containerName="glance-httpd" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.827461 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a0eb07-e443-4829-9d12-4832818ccaea" containerName="glance-httpd" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.827498 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a0eb07-e443-4829-9d12-4832818ccaea" containerName="glance-log" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.828465 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.830250 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.835838 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.953905 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-config-data\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.953945 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.953966 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-scripts\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.953990 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.954021 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5fv6\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-kube-api-access-k5fv6\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.954057 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-ceph\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.954075 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-logs\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:38 crc kubenswrapper[4871]: I1007 23:41:38.994645 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a0eb07-e443-4829-9d12-4832818ccaea" path="/var/lib/kubelet/pods/14a0eb07-e443-4829-9d12-4832818ccaea/volumes" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.007413 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.056090 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-config-data\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.056130 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.056152 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-scripts\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.056175 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.056206 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5fv6\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-kube-api-access-k5fv6\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.056242 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-ceph\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.056258 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-logs\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.056643 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-logs\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.057156 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.059779 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-ceph\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.060420 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-scripts\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.060873 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-config-data\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.069250 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.072264 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5fv6\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-kube-api-access-k5fv6\") pod \"glance-default-external-api-0\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.145400 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.645028 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.775480 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e72ca756-e9c6-47f6-9a56-ea8f8da10938","Type":"ContainerStarted","Data":"68e8bcbb6e10488cd59342a92b7fe7296e7c928fb4aa6d5c1d1a3878dff9b9c1"} Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.775652 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerName="glance-log" containerID="cri-o://232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447" gracePeriod=30 Oct 07 23:41:39 crc kubenswrapper[4871]: I1007 23:41:39.775728 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerName="glance-httpd" containerID="cri-o://af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1" gracePeriod=30 Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.380525 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.485459 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-httpd-run\") pod \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.485593 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-combined-ca-bundle\") pod \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.485623 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-ceph\") pod \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.485666 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-logs\") pod \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.485700 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shh5g\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-kube-api-access-shh5g\") pod \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.485786 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-scripts\") pod \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.485855 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-config-data\") pod \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\" (UID: \"f100e6ff-78fe-4c67-b7d8-050e4a07ad40\") " Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.485922 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f100e6ff-78fe-4c67-b7d8-050e4a07ad40" (UID: "f100e6ff-78fe-4c67-b7d8-050e4a07ad40"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.486105 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-logs" (OuterVolumeSpecName: "logs") pod "f100e6ff-78fe-4c67-b7d8-050e4a07ad40" (UID: "f100e6ff-78fe-4c67-b7d8-050e4a07ad40"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.486272 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.486289 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.490744 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-ceph" (OuterVolumeSpecName: "ceph") pod "f100e6ff-78fe-4c67-b7d8-050e4a07ad40" (UID: "f100e6ff-78fe-4c67-b7d8-050e4a07ad40"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.492610 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-kube-api-access-shh5g" (OuterVolumeSpecName: "kube-api-access-shh5g") pod "f100e6ff-78fe-4c67-b7d8-050e4a07ad40" (UID: "f100e6ff-78fe-4c67-b7d8-050e4a07ad40"). InnerVolumeSpecName "kube-api-access-shh5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.499051 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-scripts" (OuterVolumeSpecName: "scripts") pod "f100e6ff-78fe-4c67-b7d8-050e4a07ad40" (UID: "f100e6ff-78fe-4c67-b7d8-050e4a07ad40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.517954 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f100e6ff-78fe-4c67-b7d8-050e4a07ad40" (UID: "f100e6ff-78fe-4c67-b7d8-050e4a07ad40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.528729 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-config-data" (OuterVolumeSpecName: "config-data") pod "f100e6ff-78fe-4c67-b7d8-050e4a07ad40" (UID: "f100e6ff-78fe-4c67-b7d8-050e4a07ad40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.587912 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.588247 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.588262 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.588274 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.588315 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shh5g\" (UniqueName: \"kubernetes.io/projected/f100e6ff-78fe-4c67-b7d8-050e4a07ad40-kube-api-access-shh5g\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.786082 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e72ca756-e9c6-47f6-9a56-ea8f8da10938","Type":"ContainerStarted","Data":"d0811bd3822ca6ebeb120cdc6461892d29a96b4c5c061b4fc658953ddaa0eb6d"} Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.786128 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e72ca756-e9c6-47f6-9a56-ea8f8da10938","Type":"ContainerStarted","Data":"a19d16bbb1da455b8bd1a86f2aecf55b9b15a8bfd8d48b6c2852c587c419ac82"} Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.787931 4871 generic.go:334] "Generic (PLEG): container finished" podID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerID="af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1" exitCode=0 Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.787962 4871 generic.go:334] "Generic (PLEG): container finished" podID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerID="232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447" exitCode=143 Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.787972 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.787982 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f100e6ff-78fe-4c67-b7d8-050e4a07ad40","Type":"ContainerDied","Data":"af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1"} Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.788085 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f100e6ff-78fe-4c67-b7d8-050e4a07ad40","Type":"ContainerDied","Data":"232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447"} Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.788100 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f100e6ff-78fe-4c67-b7d8-050e4a07ad40","Type":"ContainerDied","Data":"78f937ada2c2ab0c715d24c503751db691afc685f80827731ace562a15aff75d"} Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.788118 4871 scope.go:117] "RemoveContainer" containerID="af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.820569 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.820546695 podStartE2EDuration="2.820546695s" podCreationTimestamp="2025-10-07 23:41:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:41:40.815224785 +0000 UTC m=+5574.617922858" watchObservedRunningTime="2025-10-07 23:41:40.820546695 +0000 UTC m=+5574.623244768" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.840813 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.842072 4871 scope.go:117] "RemoveContainer" containerID="232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.851852 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.871013 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:41:40 crc kubenswrapper[4871]: E1007 23:41:40.871704 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerName="glance-log" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.871744 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerName="glance-log" Oct 07 23:41:40 crc kubenswrapper[4871]: E1007 23:41:40.871833 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerName="glance-httpd" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.871855 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerName="glance-httpd" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.872303 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerName="glance-log" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.872382 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" containerName="glance-httpd" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.874953 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.877059 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.888646 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.905257 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-scripts\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.907877 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.909538 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.909608 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-ceph\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.909704 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-config-data\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.909741 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-logs\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.909775 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfw7v\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-kube-api-access-nfw7v\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.918998 4871 scope.go:117] "RemoveContainer" containerID="af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1" Oct 07 23:41:40 crc kubenswrapper[4871]: E1007 23:41:40.919561 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1\": container with ID starting with af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1 not found: ID does not exist" containerID="af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.919621 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1"} err="failed to get container status \"af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1\": rpc error: code = NotFound desc = could not find container \"af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1\": container with ID starting with af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1 not found: ID does not exist" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.919668 4871 scope.go:117] "RemoveContainer" containerID="232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447" Oct 07 23:41:40 crc kubenswrapper[4871]: E1007 23:41:40.920325 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447\": container with ID starting with 232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447 not found: ID does not exist" containerID="232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.920363 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447"} err="failed to get container status \"232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447\": rpc error: code = NotFound desc = could not find container \"232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447\": container with ID starting with 232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447 not found: ID does not exist" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.920387 4871 scope.go:117] "RemoveContainer" containerID="af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.923984 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1"} err="failed to get container status \"af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1\": rpc error: code = NotFound desc = could not find container \"af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1\": container with ID starting with af07387c3a7958907dec6394394fb37fbe97b21b4a326c23c9d9f0c071256ad1 not found: ID does not exist" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.924041 4871 scope.go:117] "RemoveContainer" containerID="232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.928022 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447"} err="failed to get container status \"232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447\": rpc error: code = NotFound desc = could not find container \"232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447\": container with ID starting with 232bc572944757e35fa80072cc42e420ad439bdea10b806f7cc1e5b59d83f447 not found: ID does not exist" Oct 07 23:41:40 crc kubenswrapper[4871]: I1007 23:41:40.997372 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f100e6ff-78fe-4c67-b7d8-050e4a07ad40" path="/var/lib/kubelet/pods/f100e6ff-78fe-4c67-b7d8-050e4a07ad40/volumes" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.011576 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-scripts\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.011660 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.011687 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.011722 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-ceph\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.011773 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-config-data\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.011817 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-logs\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.011843 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfw7v\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-kube-api-access-nfw7v\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.012332 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.012690 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-logs\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.016555 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-scripts\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.016708 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.019643 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-config-data\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.030585 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfw7v\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-kube-api-access-nfw7v\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.031134 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-ceph\") pod \"glance-default-internal-api-0\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.216211 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:41 crc kubenswrapper[4871]: I1007 23:41:41.813532 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:41:41 crc kubenswrapper[4871]: W1007 23:41:41.814942 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81394084_1cfb_4031_8c08_a7d59bf4d796.slice/crio-4daa10b0354e0d7d7b646b7b051440659d228c97920a1a5aa02b179937a41ec5 WatchSource:0}: Error finding container 4daa10b0354e0d7d7b646b7b051440659d228c97920a1a5aa02b179937a41ec5: Status 404 returned error can't find the container with id 4daa10b0354e0d7d7b646b7b051440659d228c97920a1a5aa02b179937a41ec5 Oct 07 23:41:42 crc kubenswrapper[4871]: I1007 23:41:42.815290 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81394084-1cfb-4031-8c08-a7d59bf4d796","Type":"ContainerStarted","Data":"6af6940675088239622d7ba305cd96b1b2662bf063f2be6efe35de6f0e8243eb"} Oct 07 23:41:42 crc kubenswrapper[4871]: I1007 23:41:42.815618 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81394084-1cfb-4031-8c08-a7d59bf4d796","Type":"ContainerStarted","Data":"4daa10b0354e0d7d7b646b7b051440659d228c97920a1a5aa02b179937a41ec5"} Oct 07 23:41:43 crc kubenswrapper[4871]: I1007 23:41:43.844888 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81394084-1cfb-4031-8c08-a7d59bf4d796","Type":"ContainerStarted","Data":"bb9a8d3c4ec333ea45b6256627673301f12cf3af1bfea730f986e0ec05b71905"} Oct 07 23:41:43 crc kubenswrapper[4871]: I1007 23:41:43.887368 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.8873462659999998 podStartE2EDuration="3.887346266s" podCreationTimestamp="2025-10-07 23:41:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:41:43.875284798 +0000 UTC m=+5577.677982911" watchObservedRunningTime="2025-10-07 23:41:43.887346266 +0000 UTC m=+5577.690044349" Oct 07 23:41:45 crc kubenswrapper[4871]: I1007 23:41:45.495056 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:41:45 crc kubenswrapper[4871]: I1007 23:41:45.579966 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b7d74c6cf-5w7gm"] Oct 07 23:41:45 crc kubenswrapper[4871]: I1007 23:41:45.580268 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" podUID="e228a39c-1929-416e-a1ab-15c66c6eb329" containerName="dnsmasq-dns" containerID="cri-o://0e19ce0ccba6c04e71df39ba8e9b1de79f106ec501428817450cb55b52a9bd24" gracePeriod=10 Oct 07 23:41:45 crc kubenswrapper[4871]: I1007 23:41:45.867835 4871 generic.go:334] "Generic (PLEG): container finished" podID="e228a39c-1929-416e-a1ab-15c66c6eb329" containerID="0e19ce0ccba6c04e71df39ba8e9b1de79f106ec501428817450cb55b52a9bd24" exitCode=0 Oct 07 23:41:45 crc kubenswrapper[4871]: I1007 23:41:45.867964 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" event={"ID":"e228a39c-1929-416e-a1ab-15c66c6eb329","Type":"ContainerDied","Data":"0e19ce0ccba6c04e71df39ba8e9b1de79f106ec501428817450cb55b52a9bd24"} Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.074764 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.123415 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wshq5\" (UniqueName: \"kubernetes.io/projected/e228a39c-1929-416e-a1ab-15c66c6eb329-kube-api-access-wshq5\") pod \"e228a39c-1929-416e-a1ab-15c66c6eb329\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.123535 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-nb\") pod \"e228a39c-1929-416e-a1ab-15c66c6eb329\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.123600 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-config\") pod \"e228a39c-1929-416e-a1ab-15c66c6eb329\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.123640 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-dns-svc\") pod \"e228a39c-1929-416e-a1ab-15c66c6eb329\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.123682 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-sb\") pod \"e228a39c-1929-416e-a1ab-15c66c6eb329\" (UID: \"e228a39c-1929-416e-a1ab-15c66c6eb329\") " Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.129424 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e228a39c-1929-416e-a1ab-15c66c6eb329-kube-api-access-wshq5" (OuterVolumeSpecName: "kube-api-access-wshq5") pod "e228a39c-1929-416e-a1ab-15c66c6eb329" (UID: "e228a39c-1929-416e-a1ab-15c66c6eb329"). InnerVolumeSpecName "kube-api-access-wshq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.175193 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-config" (OuterVolumeSpecName: "config") pod "e228a39c-1929-416e-a1ab-15c66c6eb329" (UID: "e228a39c-1929-416e-a1ab-15c66c6eb329"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.175669 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e228a39c-1929-416e-a1ab-15c66c6eb329" (UID: "e228a39c-1929-416e-a1ab-15c66c6eb329"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.180771 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e228a39c-1929-416e-a1ab-15c66c6eb329" (UID: "e228a39c-1929-416e-a1ab-15c66c6eb329"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.184743 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e228a39c-1929-416e-a1ab-15c66c6eb329" (UID: "e228a39c-1929-416e-a1ab-15c66c6eb329"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.226003 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.226052 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.226065 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.226077 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e228a39c-1929-416e-a1ab-15c66c6eb329-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.226091 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wshq5\" (UniqueName: \"kubernetes.io/projected/e228a39c-1929-416e-a1ab-15c66c6eb329-kube-api-access-wshq5\") on node \"crc\" DevicePath \"\"" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.902638 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" event={"ID":"e228a39c-1929-416e-a1ab-15c66c6eb329","Type":"ContainerDied","Data":"34bfb90b579acf96b0aabb3774119f95fad3689e68aa7a12bcf494c5c09b54e9"} Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.902783 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b7d74c6cf-5w7gm" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.902751 4871 scope.go:117] "RemoveContainer" containerID="0e19ce0ccba6c04e71df39ba8e9b1de79f106ec501428817450cb55b52a9bd24" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.954474 4871 scope.go:117] "RemoveContainer" containerID="656bb477211bd3649b1335ee2e4242b1ec560e2171baf90587abca189766b510" Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.964709 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b7d74c6cf-5w7gm"] Oct 07 23:41:46 crc kubenswrapper[4871]: I1007 23:41:46.974465 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b7d74c6cf-5w7gm"] Oct 07 23:41:47 crc kubenswrapper[4871]: I1007 23:41:47.004545 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e228a39c-1929-416e-a1ab-15c66c6eb329" path="/var/lib/kubelet/pods/e228a39c-1929-416e-a1ab-15c66c6eb329/volumes" Oct 07 23:41:49 crc kubenswrapper[4871]: I1007 23:41:49.146775 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 23:41:49 crc kubenswrapper[4871]: I1007 23:41:49.147122 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 23:41:49 crc kubenswrapper[4871]: I1007 23:41:49.190055 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 23:41:49 crc kubenswrapper[4871]: I1007 23:41:49.223355 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 23:41:49 crc kubenswrapper[4871]: I1007 23:41:49.933057 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 23:41:49 crc kubenswrapper[4871]: I1007 23:41:49.933460 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 23:41:51 crc kubenswrapper[4871]: I1007 23:41:51.217585 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:51 crc kubenswrapper[4871]: I1007 23:41:51.217675 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:51 crc kubenswrapper[4871]: I1007 23:41:51.264407 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:51 crc kubenswrapper[4871]: I1007 23:41:51.276634 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:51 crc kubenswrapper[4871]: I1007 23:41:51.837305 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 23:41:51 crc kubenswrapper[4871]: I1007 23:41:51.838542 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 23:41:51 crc kubenswrapper[4871]: I1007 23:41:51.960753 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:51 crc kubenswrapper[4871]: I1007 23:41:51.960874 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:53 crc kubenswrapper[4871]: I1007 23:41:53.832185 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:53 crc kubenswrapper[4871]: I1007 23:41:53.835110 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 23:41:57 crc kubenswrapper[4871]: I1007 23:41:57.176096 4871 scope.go:117] "RemoveContainer" containerID="f1c3dc602ed9c44b90b1e1a8c12ad3a64d6d3fed18b9726a5c3e5cf945ef6447" Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.372963 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wws9x"] Oct 07 23:42:00 crc kubenswrapper[4871]: E1007 23:42:00.373861 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e228a39c-1929-416e-a1ab-15c66c6eb329" containerName="init" Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.373877 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e228a39c-1929-416e-a1ab-15c66c6eb329" containerName="init" Oct 07 23:42:00 crc kubenswrapper[4871]: E1007 23:42:00.373906 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e228a39c-1929-416e-a1ab-15c66c6eb329" containerName="dnsmasq-dns" Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.373913 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e228a39c-1929-416e-a1ab-15c66c6eb329" containerName="dnsmasq-dns" Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.374108 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e228a39c-1929-416e-a1ab-15c66c6eb329" containerName="dnsmasq-dns" Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.374760 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wws9x" Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.390015 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wws9x"] Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.422311 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z92qk\" (UniqueName: \"kubernetes.io/projected/11b57069-5879-4ffe-9574-70103d3e0f58-kube-api-access-z92qk\") pod \"placement-db-create-wws9x\" (UID: \"11b57069-5879-4ffe-9574-70103d3e0f58\") " pod="openstack/placement-db-create-wws9x" Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.523740 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z92qk\" (UniqueName: \"kubernetes.io/projected/11b57069-5879-4ffe-9574-70103d3e0f58-kube-api-access-z92qk\") pod \"placement-db-create-wws9x\" (UID: \"11b57069-5879-4ffe-9574-70103d3e0f58\") " pod="openstack/placement-db-create-wws9x" Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.540543 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z92qk\" (UniqueName: \"kubernetes.io/projected/11b57069-5879-4ffe-9574-70103d3e0f58-kube-api-access-z92qk\") pod \"placement-db-create-wws9x\" (UID: \"11b57069-5879-4ffe-9574-70103d3e0f58\") " pod="openstack/placement-db-create-wws9x" Oct 07 23:42:00 crc kubenswrapper[4871]: I1007 23:42:00.692292 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wws9x" Oct 07 23:42:01 crc kubenswrapper[4871]: I1007 23:42:01.189007 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wws9x"] Oct 07 23:42:02 crc kubenswrapper[4871]: I1007 23:42:02.059323 4871 generic.go:334] "Generic (PLEG): container finished" podID="11b57069-5879-4ffe-9574-70103d3e0f58" containerID="c933ceb6c23a048767fe87f97d851a5718a6bac3cda3abfa6cbd7c3bff7a0d87" exitCode=0 Oct 07 23:42:02 crc kubenswrapper[4871]: I1007 23:42:02.059581 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wws9x" event={"ID":"11b57069-5879-4ffe-9574-70103d3e0f58","Type":"ContainerDied","Data":"c933ceb6c23a048767fe87f97d851a5718a6bac3cda3abfa6cbd7c3bff7a0d87"} Oct 07 23:42:02 crc kubenswrapper[4871]: I1007 23:42:02.059663 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wws9x" event={"ID":"11b57069-5879-4ffe-9574-70103d3e0f58","Type":"ContainerStarted","Data":"aaa8dd0f4925890ca712186eb8d0b5c6e1b08a0dfe91a44c04f3c81a0b5330ed"} Oct 07 23:42:03 crc kubenswrapper[4871]: I1007 23:42:03.407220 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wws9x" Oct 07 23:42:03 crc kubenswrapper[4871]: I1007 23:42:03.480594 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z92qk\" (UniqueName: \"kubernetes.io/projected/11b57069-5879-4ffe-9574-70103d3e0f58-kube-api-access-z92qk\") pod \"11b57069-5879-4ffe-9574-70103d3e0f58\" (UID: \"11b57069-5879-4ffe-9574-70103d3e0f58\") " Oct 07 23:42:03 crc kubenswrapper[4871]: I1007 23:42:03.489440 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b57069-5879-4ffe-9574-70103d3e0f58-kube-api-access-z92qk" (OuterVolumeSpecName: "kube-api-access-z92qk") pod "11b57069-5879-4ffe-9574-70103d3e0f58" (UID: "11b57069-5879-4ffe-9574-70103d3e0f58"). InnerVolumeSpecName "kube-api-access-z92qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:42:03 crc kubenswrapper[4871]: I1007 23:42:03.584056 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z92qk\" (UniqueName: \"kubernetes.io/projected/11b57069-5879-4ffe-9574-70103d3e0f58-kube-api-access-z92qk\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:04 crc kubenswrapper[4871]: I1007 23:42:04.080629 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wws9x" event={"ID":"11b57069-5879-4ffe-9574-70103d3e0f58","Type":"ContainerDied","Data":"aaa8dd0f4925890ca712186eb8d0b5c6e1b08a0dfe91a44c04f3c81a0b5330ed"} Oct 07 23:42:04 crc kubenswrapper[4871]: I1007 23:42:04.080681 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aaa8dd0f4925890ca712186eb8d0b5c6e1b08a0dfe91a44c04f3c81a0b5330ed" Oct 07 23:42:04 crc kubenswrapper[4871]: I1007 23:42:04.080730 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wws9x" Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.561557 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-3b90-account-create-rdbvt"] Oct 07 23:42:10 crc kubenswrapper[4871]: E1007 23:42:10.563428 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b57069-5879-4ffe-9574-70103d3e0f58" containerName="mariadb-database-create" Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.563463 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b57069-5879-4ffe-9574-70103d3e0f58" containerName="mariadb-database-create" Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.563908 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b57069-5879-4ffe-9574-70103d3e0f58" containerName="mariadb-database-create" Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.564950 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3b90-account-create-rdbvt" Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.568521 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.574953 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3b90-account-create-rdbvt"] Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.633068 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqp7q\" (UniqueName: \"kubernetes.io/projected/b5d3c2db-5aee-4904-8bc7-50c0b556eceb-kube-api-access-vqp7q\") pod \"placement-3b90-account-create-rdbvt\" (UID: \"b5d3c2db-5aee-4904-8bc7-50c0b556eceb\") " pod="openstack/placement-3b90-account-create-rdbvt" Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.735038 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqp7q\" (UniqueName: \"kubernetes.io/projected/b5d3c2db-5aee-4904-8bc7-50c0b556eceb-kube-api-access-vqp7q\") pod \"placement-3b90-account-create-rdbvt\" (UID: \"b5d3c2db-5aee-4904-8bc7-50c0b556eceb\") " pod="openstack/placement-3b90-account-create-rdbvt" Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.768176 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqp7q\" (UniqueName: \"kubernetes.io/projected/b5d3c2db-5aee-4904-8bc7-50c0b556eceb-kube-api-access-vqp7q\") pod \"placement-3b90-account-create-rdbvt\" (UID: \"b5d3c2db-5aee-4904-8bc7-50c0b556eceb\") " pod="openstack/placement-3b90-account-create-rdbvt" Oct 07 23:42:10 crc kubenswrapper[4871]: I1007 23:42:10.901349 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3b90-account-create-rdbvt" Oct 07 23:42:11 crc kubenswrapper[4871]: I1007 23:42:11.339858 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3b90-account-create-rdbvt"] Oct 07 23:42:12 crc kubenswrapper[4871]: I1007 23:42:12.173950 4871 generic.go:334] "Generic (PLEG): container finished" podID="b5d3c2db-5aee-4904-8bc7-50c0b556eceb" containerID="2bca1fdad9767568e53e746fac1cec950cf85da925f66f971a0e074a54250a6a" exitCode=0 Oct 07 23:42:12 crc kubenswrapper[4871]: I1007 23:42:12.174407 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3b90-account-create-rdbvt" event={"ID":"b5d3c2db-5aee-4904-8bc7-50c0b556eceb","Type":"ContainerDied","Data":"2bca1fdad9767568e53e746fac1cec950cf85da925f66f971a0e074a54250a6a"} Oct 07 23:42:12 crc kubenswrapper[4871]: I1007 23:42:12.174460 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3b90-account-create-rdbvt" event={"ID":"b5d3c2db-5aee-4904-8bc7-50c0b556eceb","Type":"ContainerStarted","Data":"7a8e7266c6282abecccdaa4d40b7e0db373cafa2b462964a3c8300c0f29db7e8"} Oct 07 23:42:13 crc kubenswrapper[4871]: I1007 23:42:13.563904 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3b90-account-create-rdbvt" Oct 07 23:42:13 crc kubenswrapper[4871]: I1007 23:42:13.581244 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqp7q\" (UniqueName: \"kubernetes.io/projected/b5d3c2db-5aee-4904-8bc7-50c0b556eceb-kube-api-access-vqp7q\") pod \"b5d3c2db-5aee-4904-8bc7-50c0b556eceb\" (UID: \"b5d3c2db-5aee-4904-8bc7-50c0b556eceb\") " Oct 07 23:42:13 crc kubenswrapper[4871]: I1007 23:42:13.629057 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d3c2db-5aee-4904-8bc7-50c0b556eceb-kube-api-access-vqp7q" (OuterVolumeSpecName: "kube-api-access-vqp7q") pod "b5d3c2db-5aee-4904-8bc7-50c0b556eceb" (UID: "b5d3c2db-5aee-4904-8bc7-50c0b556eceb"). InnerVolumeSpecName "kube-api-access-vqp7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:42:13 crc kubenswrapper[4871]: I1007 23:42:13.687553 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqp7q\" (UniqueName: \"kubernetes.io/projected/b5d3c2db-5aee-4904-8bc7-50c0b556eceb-kube-api-access-vqp7q\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:14 crc kubenswrapper[4871]: I1007 23:42:14.193108 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3b90-account-create-rdbvt" event={"ID":"b5d3c2db-5aee-4904-8bc7-50c0b556eceb","Type":"ContainerDied","Data":"7a8e7266c6282abecccdaa4d40b7e0db373cafa2b462964a3c8300c0f29db7e8"} Oct 07 23:42:14 crc kubenswrapper[4871]: I1007 23:42:14.193147 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a8e7266c6282abecccdaa4d40b7e0db373cafa2b462964a3c8300c0f29db7e8" Oct 07 23:42:14 crc kubenswrapper[4871]: I1007 23:42:14.193157 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3b90-account-create-rdbvt" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.891609 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b67b947d7-cks8k"] Oct 07 23:42:15 crc kubenswrapper[4871]: E1007 23:42:15.893446 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d3c2db-5aee-4904-8bc7-50c0b556eceb" containerName="mariadb-account-create" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.893466 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d3c2db-5aee-4904-8bc7-50c0b556eceb" containerName="mariadb-account-create" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.893640 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d3c2db-5aee-4904-8bc7-50c0b556eceb" containerName="mariadb-account-create" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.894899 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.901043 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b67b947d7-cks8k"] Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.927358 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-config\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.927399 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-sb\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.927452 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-dns-svc\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.927495 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdvwm\" (UniqueName: \"kubernetes.io/projected/00e42cab-5504-46f1-a7ad-a74f4007bfef-kube-api-access-jdvwm\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.927541 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-nb\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.940084 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-2gqc9"] Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.941573 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.948413 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2gqc9"] Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.951479 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.951670 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-s97g9" Oct 07 23:42:15 crc kubenswrapper[4871]: I1007 23:42:15.952248 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.028427 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-config\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.028467 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-sb\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.028494 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-config-data\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.028536 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-scripts\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.028554 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-dns-svc\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.028572 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-logs\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.028761 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdvwm\" (UniqueName: \"kubernetes.io/projected/00e42cab-5504-46f1-a7ad-a74f4007bfef-kube-api-access-jdvwm\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.028844 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xp9q\" (UniqueName: \"kubernetes.io/projected/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-kube-api-access-8xp9q\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.028970 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-nb\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.029016 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-combined-ca-bundle\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.029402 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-sb\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.030173 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-nb\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.029477 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-dns-svc\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.030401 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-config\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.051283 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdvwm\" (UniqueName: \"kubernetes.io/projected/00e42cab-5504-46f1-a7ad-a74f4007bfef-kube-api-access-jdvwm\") pod \"dnsmasq-dns-b67b947d7-cks8k\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.130032 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-combined-ca-bundle\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.130100 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-config-data\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.130150 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-scripts\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.130167 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-logs\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.130220 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xp9q\" (UniqueName: \"kubernetes.io/projected/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-kube-api-access-8xp9q\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.130862 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-logs\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.132844 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-scripts\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.133421 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-combined-ca-bundle\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.133917 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-config-data\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.145404 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xp9q\" (UniqueName: \"kubernetes.io/projected/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-kube-api-access-8xp9q\") pod \"placement-db-sync-2gqc9\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.225573 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.270465 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.679470 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b67b947d7-cks8k"] Oct 07 23:42:16 crc kubenswrapper[4871]: I1007 23:42:16.768029 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2gqc9"] Oct 07 23:42:16 crc kubenswrapper[4871]: W1007 23:42:16.788411 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9b014be_6d59_4c4c_a932_3d5a9efe4b12.slice/crio-6b6f800c78ea0f52f87ba92fdb18969775902380d97180df2fe65d2af7572d0d WatchSource:0}: Error finding container 6b6f800c78ea0f52f87ba92fdb18969775902380d97180df2fe65d2af7572d0d: Status 404 returned error can't find the container with id 6b6f800c78ea0f52f87ba92fdb18969775902380d97180df2fe65d2af7572d0d Oct 07 23:42:17 crc kubenswrapper[4871]: I1007 23:42:17.219845 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2gqc9" event={"ID":"e9b014be-6d59-4c4c-a932-3d5a9efe4b12","Type":"ContainerStarted","Data":"4b3968107feab7d9fc685588f70fb5321e0c37bb16c7db451fb28831edc7fa53"} Oct 07 23:42:17 crc kubenswrapper[4871]: I1007 23:42:17.220332 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2gqc9" event={"ID":"e9b014be-6d59-4c4c-a932-3d5a9efe4b12","Type":"ContainerStarted","Data":"6b6f800c78ea0f52f87ba92fdb18969775902380d97180df2fe65d2af7572d0d"} Oct 07 23:42:17 crc kubenswrapper[4871]: I1007 23:42:17.225399 4871 generic.go:334] "Generic (PLEG): container finished" podID="00e42cab-5504-46f1-a7ad-a74f4007bfef" containerID="b493fd5137e12dc4e6666e5e6543a533c297a9bf0a5c63089cf0c03c155b9869" exitCode=0 Oct 07 23:42:17 crc kubenswrapper[4871]: I1007 23:42:17.225480 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" event={"ID":"00e42cab-5504-46f1-a7ad-a74f4007bfef","Type":"ContainerDied","Data":"b493fd5137e12dc4e6666e5e6543a533c297a9bf0a5c63089cf0c03c155b9869"} Oct 07 23:42:17 crc kubenswrapper[4871]: I1007 23:42:17.225617 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" event={"ID":"00e42cab-5504-46f1-a7ad-a74f4007bfef","Type":"ContainerStarted","Data":"4e1a7528e05d9edf577bcee741a164d991f1b2bb8ba96568bbd4998296a41ed0"} Oct 07 23:42:17 crc kubenswrapper[4871]: I1007 23:42:17.248359 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-2gqc9" podStartSLOduration=2.248334702 podStartE2EDuration="2.248334702s" podCreationTimestamp="2025-10-07 23:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:42:17.237925817 +0000 UTC m=+5611.040623890" watchObservedRunningTime="2025-10-07 23:42:17.248334702 +0000 UTC m=+5611.051032785" Oct 07 23:42:18 crc kubenswrapper[4871]: I1007 23:42:18.239331 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" event={"ID":"00e42cab-5504-46f1-a7ad-a74f4007bfef","Type":"ContainerStarted","Data":"8885cb7238f5b3a02820c42f2a0544f7b5ff9f92b5de2a2fab0f68d83810b600"} Oct 07 23:42:18 crc kubenswrapper[4871]: I1007 23:42:18.265056 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" podStartSLOduration=3.265031702 podStartE2EDuration="3.265031702s" podCreationTimestamp="2025-10-07 23:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:42:18.262068374 +0000 UTC m=+5612.064766487" watchObservedRunningTime="2025-10-07 23:42:18.265031702 +0000 UTC m=+5612.067729785" Oct 07 23:42:19 crc kubenswrapper[4871]: I1007 23:42:19.250103 4871 generic.go:334] "Generic (PLEG): container finished" podID="e9b014be-6d59-4c4c-a932-3d5a9efe4b12" containerID="4b3968107feab7d9fc685588f70fb5321e0c37bb16c7db451fb28831edc7fa53" exitCode=0 Oct 07 23:42:19 crc kubenswrapper[4871]: I1007 23:42:19.250205 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2gqc9" event={"ID":"e9b014be-6d59-4c4c-a932-3d5a9efe4b12","Type":"ContainerDied","Data":"4b3968107feab7d9fc685588f70fb5321e0c37bb16c7db451fb28831edc7fa53"} Oct 07 23:42:19 crc kubenswrapper[4871]: I1007 23:42:19.251679 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.659574 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.820603 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-combined-ca-bundle\") pod \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.820769 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-scripts\") pod \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.820789 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xp9q\" (UniqueName: \"kubernetes.io/projected/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-kube-api-access-8xp9q\") pod \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.821336 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-logs\") pod \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.821393 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-config-data\") pod \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\" (UID: \"e9b014be-6d59-4c4c-a932-3d5a9efe4b12\") " Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.821870 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-logs" (OuterVolumeSpecName: "logs") pod "e9b014be-6d59-4c4c-a932-3d5a9efe4b12" (UID: "e9b014be-6d59-4c4c-a932-3d5a9efe4b12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.822187 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.832921 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-scripts" (OuterVolumeSpecName: "scripts") pod "e9b014be-6d59-4c4c-a932-3d5a9efe4b12" (UID: "e9b014be-6d59-4c4c-a932-3d5a9efe4b12"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.833011 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-kube-api-access-8xp9q" (OuterVolumeSpecName: "kube-api-access-8xp9q") pod "e9b014be-6d59-4c4c-a932-3d5a9efe4b12" (UID: "e9b014be-6d59-4c4c-a932-3d5a9efe4b12"). InnerVolumeSpecName "kube-api-access-8xp9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.847136 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9b014be-6d59-4c4c-a932-3d5a9efe4b12" (UID: "e9b014be-6d59-4c4c-a932-3d5a9efe4b12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.850272 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-config-data" (OuterVolumeSpecName: "config-data") pod "e9b014be-6d59-4c4c-a932-3d5a9efe4b12" (UID: "e9b014be-6d59-4c4c-a932-3d5a9efe4b12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.928917 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.928951 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.928985 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:20 crc kubenswrapper[4871]: I1007 23:42:20.928995 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xp9q\" (UniqueName: \"kubernetes.io/projected/e9b014be-6d59-4c4c-a932-3d5a9efe4b12-kube-api-access-8xp9q\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.290630 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2gqc9" event={"ID":"e9b014be-6d59-4c4c-a932-3d5a9efe4b12","Type":"ContainerDied","Data":"6b6f800c78ea0f52f87ba92fdb18969775902380d97180df2fe65d2af7572d0d"} Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.290864 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2gqc9" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.291206 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b6f800c78ea0f52f87ba92fdb18969775902380d97180df2fe65d2af7572d0d" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.369831 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c884b686b-r8slm"] Oct 07 23:42:21 crc kubenswrapper[4871]: E1007 23:42:21.370244 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9b014be-6d59-4c4c-a932-3d5a9efe4b12" containerName="placement-db-sync" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.370263 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9b014be-6d59-4c4c-a932-3d5a9efe4b12" containerName="placement-db-sync" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.370475 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9b014be-6d59-4c4c-a932-3d5a9efe4b12" containerName="placement-db-sync" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.371548 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.375007 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-s97g9" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.375259 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.375415 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.385717 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c884b686b-r8slm"] Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.537914 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-scripts\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.538027 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmtnc\" (UniqueName: \"kubernetes.io/projected/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-kube-api-access-wmtnc\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.538260 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-logs\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.538422 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-config-data\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.538475 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-combined-ca-bundle\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.640393 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-scripts\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.640478 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmtnc\" (UniqueName: \"kubernetes.io/projected/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-kube-api-access-wmtnc\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.640582 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-logs\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.640640 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-config-data\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.640665 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-combined-ca-bundle\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.641694 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-logs\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.645499 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-scripts\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.646789 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-combined-ca-bundle\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.647663 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-config-data\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.673986 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmtnc\" (UniqueName: \"kubernetes.io/projected/019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57-kube-api-access-wmtnc\") pod \"placement-c884b686b-r8slm\" (UID: \"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57\") " pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:21 crc kubenswrapper[4871]: I1007 23:42:21.734565 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:22 crc kubenswrapper[4871]: I1007 23:42:22.253862 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c884b686b-r8slm"] Oct 07 23:42:22 crc kubenswrapper[4871]: I1007 23:42:22.302681 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c884b686b-r8slm" event={"ID":"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57","Type":"ContainerStarted","Data":"644de0fb3613da188b2829fdaf7ac17ec91143d06fb3027ff698d4c51839d609"} Oct 07 23:42:23 crc kubenswrapper[4871]: I1007 23:42:23.316623 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c884b686b-r8slm" event={"ID":"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57","Type":"ContainerStarted","Data":"73cd83c074fd2a05f0f53bcd7cfd4f54c127f0114362369274ecefe96fcbbdc1"} Oct 07 23:42:23 crc kubenswrapper[4871]: I1007 23:42:23.317047 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:23 crc kubenswrapper[4871]: I1007 23:42:23.317064 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c884b686b-r8slm" event={"ID":"019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57","Type":"ContainerStarted","Data":"b9b5de1c8f2fb8a5e22c578433565edd675b1e6ecbb8f4945f2e87f96b3dea1f"} Oct 07 23:42:23 crc kubenswrapper[4871]: I1007 23:42:23.317080 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:23 crc kubenswrapper[4871]: I1007 23:42:23.351559 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-c884b686b-r8slm" podStartSLOduration=2.351523431 podStartE2EDuration="2.351523431s" podCreationTimestamp="2025-10-07 23:42:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:42:23.34658739 +0000 UTC m=+5617.149285463" watchObservedRunningTime="2025-10-07 23:42:23.351523431 +0000 UTC m=+5617.154221554" Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.227037 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.314056 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bcbfc77f9-sljjb"] Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.314417 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" podUID="336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" containerName="dnsmasq-dns" containerID="cri-o://16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad" gracePeriod=10 Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.784337 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.942566 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-sb\") pod \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.942764 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-config\") pod \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.943164 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-dns-svc\") pod \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.943257 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgsxk\" (UniqueName: \"kubernetes.io/projected/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-kube-api-access-bgsxk\") pod \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.943278 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-nb\") pod \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\" (UID: \"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3\") " Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.951701 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-kube-api-access-bgsxk" (OuterVolumeSpecName: "kube-api-access-bgsxk") pod "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" (UID: "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3"). InnerVolumeSpecName "kube-api-access-bgsxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.982234 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" (UID: "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:42:26 crc kubenswrapper[4871]: I1007 23:42:26.991374 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" (UID: "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.006174 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-config" (OuterVolumeSpecName: "config") pod "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" (UID: "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.031845 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" (UID: "336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.045570 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgsxk\" (UniqueName: \"kubernetes.io/projected/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-kube-api-access-bgsxk\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.045600 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.045610 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.045619 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.045627 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.384787 4871 generic.go:334] "Generic (PLEG): container finished" podID="336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" containerID="16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad" exitCode=0 Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.384872 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" event={"ID":"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3","Type":"ContainerDied","Data":"16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad"} Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.384956 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" event={"ID":"336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3","Type":"ContainerDied","Data":"c3371ab2a17405621a065f1f2ba27a462730a494d5647f27b77ca405810d75b0"} Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.384981 4871 scope.go:117] "RemoveContainer" containerID="16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.385941 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bcbfc77f9-sljjb" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.428919 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bcbfc77f9-sljjb"] Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.435390 4871 scope.go:117] "RemoveContainer" containerID="8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.439472 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bcbfc77f9-sljjb"] Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.473482 4871 scope.go:117] "RemoveContainer" containerID="16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad" Oct 07 23:42:27 crc kubenswrapper[4871]: E1007 23:42:27.473915 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad\": container with ID starting with 16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad not found: ID does not exist" containerID="16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.473956 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad"} err="failed to get container status \"16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad\": rpc error: code = NotFound desc = could not find container \"16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad\": container with ID starting with 16c0d6571ef9194bc93955b02d11de02b8bb3c00c3a52797e2caa7bab65605ad not found: ID does not exist" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.473986 4871 scope.go:117] "RemoveContainer" containerID="8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf" Oct 07 23:42:27 crc kubenswrapper[4871]: E1007 23:42:27.474329 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf\": container with ID starting with 8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf not found: ID does not exist" containerID="8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf" Oct 07 23:42:27 crc kubenswrapper[4871]: I1007 23:42:27.474364 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf"} err="failed to get container status \"8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf\": rpc error: code = NotFound desc = could not find container \"8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf\": container with ID starting with 8a20552b5b00ab18ed632d6a7a2e287bddc90b8ece98d0f951b02a3a952eebaf not found: ID does not exist" Oct 07 23:42:28 crc kubenswrapper[4871]: I1007 23:42:28.993581 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" path="/var/lib/kubelet/pods/336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3/volumes" Oct 07 23:42:35 crc kubenswrapper[4871]: I1007 23:42:35.512023 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:42:35 crc kubenswrapper[4871]: I1007 23:42:35.512902 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:42:52 crc kubenswrapper[4871]: I1007 23:42:52.741453 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:52 crc kubenswrapper[4871]: I1007 23:42:52.743131 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c884b686b-r8slm" Oct 07 23:42:57 crc kubenswrapper[4871]: I1007 23:42:57.301460 4871 scope.go:117] "RemoveContainer" containerID="81f06d6978c0d263f069798afa33d7e83748cc970921e9e76abe61a4d3b7e061" Oct 07 23:43:05 crc kubenswrapper[4871]: I1007 23:43:05.512598 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:43:05 crc kubenswrapper[4871]: I1007 23:43:05.513086 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.654752 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-pdvjn"] Oct 07 23:43:13 crc kubenswrapper[4871]: E1007 23:43:13.655514 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" containerName="dnsmasq-dns" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.655526 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" containerName="dnsmasq-dns" Oct 07 23:43:13 crc kubenswrapper[4871]: E1007 23:43:13.655547 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" containerName="init" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.655554 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" containerName="init" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.655707 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="336b82c8-eaa3-4af8-b687-bb4ae8dfe6f3" containerName="dnsmasq-dns" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.656252 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pdvjn" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.677056 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pdvjn"] Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.739603 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-4hfxw"] Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.740893 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4hfxw" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.753597 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4hfxw"] Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.795475 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6vw9\" (UniqueName: \"kubernetes.io/projected/b7141cb8-92b3-4b46-b0f7-29fb0b48d479-kube-api-access-f6vw9\") pod \"nova-api-db-create-pdvjn\" (UID: \"b7141cb8-92b3-4b46-b0f7-29fb0b48d479\") " pod="openstack/nova-api-db-create-pdvjn" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.853323 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-q6fp5"] Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.861407 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-q6fp5" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.865743 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-q6fp5"] Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.896784 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6vw9\" (UniqueName: \"kubernetes.io/projected/b7141cb8-92b3-4b46-b0f7-29fb0b48d479-kube-api-access-f6vw9\") pod \"nova-api-db-create-pdvjn\" (UID: \"b7141cb8-92b3-4b46-b0f7-29fb0b48d479\") " pod="openstack/nova-api-db-create-pdvjn" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.896870 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbcwx\" (UniqueName: \"kubernetes.io/projected/5281c6fc-e54b-4eba-b710-b62ea4e2a9de-kube-api-access-vbcwx\") pod \"nova-cell0-db-create-4hfxw\" (UID: \"5281c6fc-e54b-4eba-b710-b62ea4e2a9de\") " pod="openstack/nova-cell0-db-create-4hfxw" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.915923 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6vw9\" (UniqueName: \"kubernetes.io/projected/b7141cb8-92b3-4b46-b0f7-29fb0b48d479-kube-api-access-f6vw9\") pod \"nova-api-db-create-pdvjn\" (UID: \"b7141cb8-92b3-4b46-b0f7-29fb0b48d479\") " pod="openstack/nova-api-db-create-pdvjn" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.976317 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pdvjn" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.998572 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbcwx\" (UniqueName: \"kubernetes.io/projected/5281c6fc-e54b-4eba-b710-b62ea4e2a9de-kube-api-access-vbcwx\") pod \"nova-cell0-db-create-4hfxw\" (UID: \"5281c6fc-e54b-4eba-b710-b62ea4e2a9de\") " pod="openstack/nova-cell0-db-create-4hfxw" Oct 07 23:43:13 crc kubenswrapper[4871]: I1007 23:43:13.998748 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rqqf\" (UniqueName: \"kubernetes.io/projected/49092041-b1e7-422f-9e09-b398fb63a548-kube-api-access-6rqqf\") pod \"nova-cell1-db-create-q6fp5\" (UID: \"49092041-b1e7-422f-9e09-b398fb63a548\") " pod="openstack/nova-cell1-db-create-q6fp5" Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.017387 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbcwx\" (UniqueName: \"kubernetes.io/projected/5281c6fc-e54b-4eba-b710-b62ea4e2a9de-kube-api-access-vbcwx\") pod \"nova-cell0-db-create-4hfxw\" (UID: \"5281c6fc-e54b-4eba-b710-b62ea4e2a9de\") " pod="openstack/nova-cell0-db-create-4hfxw" Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.062155 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4hfxw" Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.100085 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rqqf\" (UniqueName: \"kubernetes.io/projected/49092041-b1e7-422f-9e09-b398fb63a548-kube-api-access-6rqqf\") pod \"nova-cell1-db-create-q6fp5\" (UID: \"49092041-b1e7-422f-9e09-b398fb63a548\") " pod="openstack/nova-cell1-db-create-q6fp5" Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.116864 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rqqf\" (UniqueName: \"kubernetes.io/projected/49092041-b1e7-422f-9e09-b398fb63a548-kube-api-access-6rqqf\") pod \"nova-cell1-db-create-q6fp5\" (UID: \"49092041-b1e7-422f-9e09-b398fb63a548\") " pod="openstack/nova-cell1-db-create-q6fp5" Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.176661 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-q6fp5" Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.460984 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pdvjn"] Oct 07 23:43:14 crc kubenswrapper[4871]: W1007 23:43:14.467037 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7141cb8_92b3_4b46_b0f7_29fb0b48d479.slice/crio-c8a1c00d90821612a927eef06f646ca0a95e55be18fa28e1e549cd764b8b0108 WatchSource:0}: Error finding container c8a1c00d90821612a927eef06f646ca0a95e55be18fa28e1e549cd764b8b0108: Status 404 returned error can't find the container with id c8a1c00d90821612a927eef06f646ca0a95e55be18fa28e1e549cd764b8b0108 Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.587688 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4hfxw"] Oct 07 23:43:14 crc kubenswrapper[4871]: W1007 23:43:14.591999 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5281c6fc_e54b_4eba_b710_b62ea4e2a9de.slice/crio-76ef1a05e59a01c28e9b588770957414ec46aec77528f083b929b5065c56b827 WatchSource:0}: Error finding container 76ef1a05e59a01c28e9b588770957414ec46aec77528f083b929b5065c56b827: Status 404 returned error can't find the container with id 76ef1a05e59a01c28e9b588770957414ec46aec77528f083b929b5065c56b827 Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.897480 4871 generic.go:334] "Generic (PLEG): container finished" podID="5281c6fc-e54b-4eba-b710-b62ea4e2a9de" containerID="3af5e9f9da3b45e938d3a8b0680b953996e731a5d83a7608ac7bd49fd9cfba11" exitCode=0 Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.898184 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4hfxw" event={"ID":"5281c6fc-e54b-4eba-b710-b62ea4e2a9de","Type":"ContainerDied","Data":"3af5e9f9da3b45e938d3a8b0680b953996e731a5d83a7608ac7bd49fd9cfba11"} Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.898248 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4hfxw" event={"ID":"5281c6fc-e54b-4eba-b710-b62ea4e2a9de","Type":"ContainerStarted","Data":"76ef1a05e59a01c28e9b588770957414ec46aec77528f083b929b5065c56b827"} Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.903143 4871 generic.go:334] "Generic (PLEG): container finished" podID="b7141cb8-92b3-4b46-b0f7-29fb0b48d479" containerID="774c9bb65e51aeb69df62b157c02aa8bbd666899351e7f478ed28c8073822c62" exitCode=0 Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.903193 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pdvjn" event={"ID":"b7141cb8-92b3-4b46-b0f7-29fb0b48d479","Type":"ContainerDied","Data":"774c9bb65e51aeb69df62b157c02aa8bbd666899351e7f478ed28c8073822c62"} Oct 07 23:43:14 crc kubenswrapper[4871]: I1007 23:43:14.903230 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pdvjn" event={"ID":"b7141cb8-92b3-4b46-b0f7-29fb0b48d479","Type":"ContainerStarted","Data":"c8a1c00d90821612a927eef06f646ca0a95e55be18fa28e1e549cd764b8b0108"} Oct 07 23:43:15 crc kubenswrapper[4871]: I1007 23:43:15.389118 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-q6fp5"] Oct 07 23:43:15 crc kubenswrapper[4871]: I1007 23:43:15.917435 4871 generic.go:334] "Generic (PLEG): container finished" podID="49092041-b1e7-422f-9e09-b398fb63a548" containerID="f48675f0e35395a85e4947dc21ed550901f70bd64fb64f4b4fd549f4b6ee63d1" exitCode=0 Oct 07 23:43:15 crc kubenswrapper[4871]: I1007 23:43:15.917526 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-q6fp5" event={"ID":"49092041-b1e7-422f-9e09-b398fb63a548","Type":"ContainerDied","Data":"f48675f0e35395a85e4947dc21ed550901f70bd64fb64f4b4fd549f4b6ee63d1"} Oct 07 23:43:15 crc kubenswrapper[4871]: I1007 23:43:15.917590 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-q6fp5" event={"ID":"49092041-b1e7-422f-9e09-b398fb63a548","Type":"ContainerStarted","Data":"2586cd438e41099542b343b199d5b87d4cbb9e09bb067dc4cf65eb2f22b07d9d"} Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.348025 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4hfxw" Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.352297 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pdvjn" Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.460599 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6vw9\" (UniqueName: \"kubernetes.io/projected/b7141cb8-92b3-4b46-b0f7-29fb0b48d479-kube-api-access-f6vw9\") pod \"b7141cb8-92b3-4b46-b0f7-29fb0b48d479\" (UID: \"b7141cb8-92b3-4b46-b0f7-29fb0b48d479\") " Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.460704 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbcwx\" (UniqueName: \"kubernetes.io/projected/5281c6fc-e54b-4eba-b710-b62ea4e2a9de-kube-api-access-vbcwx\") pod \"5281c6fc-e54b-4eba-b710-b62ea4e2a9de\" (UID: \"5281c6fc-e54b-4eba-b710-b62ea4e2a9de\") " Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.466926 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7141cb8-92b3-4b46-b0f7-29fb0b48d479-kube-api-access-f6vw9" (OuterVolumeSpecName: "kube-api-access-f6vw9") pod "b7141cb8-92b3-4b46-b0f7-29fb0b48d479" (UID: "b7141cb8-92b3-4b46-b0f7-29fb0b48d479"). InnerVolumeSpecName "kube-api-access-f6vw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.468440 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5281c6fc-e54b-4eba-b710-b62ea4e2a9de-kube-api-access-vbcwx" (OuterVolumeSpecName: "kube-api-access-vbcwx") pod "5281c6fc-e54b-4eba-b710-b62ea4e2a9de" (UID: "5281c6fc-e54b-4eba-b710-b62ea4e2a9de"). InnerVolumeSpecName "kube-api-access-vbcwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.562270 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6vw9\" (UniqueName: \"kubernetes.io/projected/b7141cb8-92b3-4b46-b0f7-29fb0b48d479-kube-api-access-f6vw9\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.562298 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbcwx\" (UniqueName: \"kubernetes.io/projected/5281c6fc-e54b-4eba-b710-b62ea4e2a9de-kube-api-access-vbcwx\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.932944 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pdvjn" event={"ID":"b7141cb8-92b3-4b46-b0f7-29fb0b48d479","Type":"ContainerDied","Data":"c8a1c00d90821612a927eef06f646ca0a95e55be18fa28e1e549cd764b8b0108"} Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.932984 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8a1c00d90821612a927eef06f646ca0a95e55be18fa28e1e549cd764b8b0108" Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.933015 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pdvjn" Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.937250 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4hfxw" Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.937472 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4hfxw" event={"ID":"5281c6fc-e54b-4eba-b710-b62ea4e2a9de","Type":"ContainerDied","Data":"76ef1a05e59a01c28e9b588770957414ec46aec77528f083b929b5065c56b827"} Oct 07 23:43:16 crc kubenswrapper[4871]: I1007 23:43:16.937505 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76ef1a05e59a01c28e9b588770957414ec46aec77528f083b929b5065c56b827" Oct 07 23:43:17 crc kubenswrapper[4871]: I1007 23:43:17.218727 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-q6fp5" Oct 07 23:43:17 crc kubenswrapper[4871]: I1007 23:43:17.290579 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rqqf\" (UniqueName: \"kubernetes.io/projected/49092041-b1e7-422f-9e09-b398fb63a548-kube-api-access-6rqqf\") pod \"49092041-b1e7-422f-9e09-b398fb63a548\" (UID: \"49092041-b1e7-422f-9e09-b398fb63a548\") " Oct 07 23:43:17 crc kubenswrapper[4871]: I1007 23:43:17.294214 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49092041-b1e7-422f-9e09-b398fb63a548-kube-api-access-6rqqf" (OuterVolumeSpecName: "kube-api-access-6rqqf") pod "49092041-b1e7-422f-9e09-b398fb63a548" (UID: "49092041-b1e7-422f-9e09-b398fb63a548"). InnerVolumeSpecName "kube-api-access-6rqqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:17 crc kubenswrapper[4871]: I1007 23:43:17.392780 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rqqf\" (UniqueName: \"kubernetes.io/projected/49092041-b1e7-422f-9e09-b398fb63a548-kube-api-access-6rqqf\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:17 crc kubenswrapper[4871]: I1007 23:43:17.947706 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-q6fp5" event={"ID":"49092041-b1e7-422f-9e09-b398fb63a548","Type":"ContainerDied","Data":"2586cd438e41099542b343b199d5b87d4cbb9e09bb067dc4cf65eb2f22b07d9d"} Oct 07 23:43:17 crc kubenswrapper[4871]: I1007 23:43:17.947753 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2586cd438e41099542b343b199d5b87d4cbb9e09bb067dc4cf65eb2f22b07d9d" Oct 07 23:43:17 crc kubenswrapper[4871]: I1007 23:43:17.947861 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-q6fp5" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.899581 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-7720-account-create-78zkj"] Oct 07 23:43:23 crc kubenswrapper[4871]: E1007 23:43:23.900562 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5281c6fc-e54b-4eba-b710-b62ea4e2a9de" containerName="mariadb-database-create" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.900580 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5281c6fc-e54b-4eba-b710-b62ea4e2a9de" containerName="mariadb-database-create" Oct 07 23:43:23 crc kubenswrapper[4871]: E1007 23:43:23.900593 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7141cb8-92b3-4b46-b0f7-29fb0b48d479" containerName="mariadb-database-create" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.900602 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7141cb8-92b3-4b46-b0f7-29fb0b48d479" containerName="mariadb-database-create" Oct 07 23:43:23 crc kubenswrapper[4871]: E1007 23:43:23.900627 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49092041-b1e7-422f-9e09-b398fb63a548" containerName="mariadb-database-create" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.900635 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="49092041-b1e7-422f-9e09-b398fb63a548" containerName="mariadb-database-create" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.900864 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7141cb8-92b3-4b46-b0f7-29fb0b48d479" containerName="mariadb-database-create" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.900887 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="49092041-b1e7-422f-9e09-b398fb63a548" containerName="mariadb-database-create" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.900911 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5281c6fc-e54b-4eba-b710-b62ea4e2a9de" containerName="mariadb-database-create" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.901561 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7720-account-create-78zkj" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.903760 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 07 23:43:23 crc kubenswrapper[4871]: I1007 23:43:23.954432 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7720-account-create-78zkj"] Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.040739 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz4fr\" (UniqueName: \"kubernetes.io/projected/833593af-531f-4b85-ab9f-56d6a657de34-kube-api-access-gz4fr\") pod \"nova-api-7720-account-create-78zkj\" (UID: \"833593af-531f-4b85-ab9f-56d6a657de34\") " pod="openstack/nova-api-7720-account-create-78zkj" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.103906 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-73a7-account-create-j6h4l"] Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.105232 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-73a7-account-create-j6h4l" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.107738 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.120619 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-73a7-account-create-j6h4l"] Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.142623 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz4fr\" (UniqueName: \"kubernetes.io/projected/833593af-531f-4b85-ab9f-56d6a657de34-kube-api-access-gz4fr\") pod \"nova-api-7720-account-create-78zkj\" (UID: \"833593af-531f-4b85-ab9f-56d6a657de34\") " pod="openstack/nova-api-7720-account-create-78zkj" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.180138 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz4fr\" (UniqueName: \"kubernetes.io/projected/833593af-531f-4b85-ab9f-56d6a657de34-kube-api-access-gz4fr\") pod \"nova-api-7720-account-create-78zkj\" (UID: \"833593af-531f-4b85-ab9f-56d6a657de34\") " pod="openstack/nova-api-7720-account-create-78zkj" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.227234 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7720-account-create-78zkj" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.244040 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxlzg\" (UniqueName: \"kubernetes.io/projected/f2ab3195-0fe3-46cf-b07b-ae48b41273e5-kube-api-access-lxlzg\") pod \"nova-cell0-73a7-account-create-j6h4l\" (UID: \"f2ab3195-0fe3-46cf-b07b-ae48b41273e5\") " pod="openstack/nova-cell0-73a7-account-create-j6h4l" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.294381 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-77da-account-create-wr7bc"] Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.295633 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-77da-account-create-wr7bc" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.298081 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.307177 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-77da-account-create-wr7bc"] Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.346530 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxlzg\" (UniqueName: \"kubernetes.io/projected/f2ab3195-0fe3-46cf-b07b-ae48b41273e5-kube-api-access-lxlzg\") pod \"nova-cell0-73a7-account-create-j6h4l\" (UID: \"f2ab3195-0fe3-46cf-b07b-ae48b41273e5\") " pod="openstack/nova-cell0-73a7-account-create-j6h4l" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.364443 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxlzg\" (UniqueName: \"kubernetes.io/projected/f2ab3195-0fe3-46cf-b07b-ae48b41273e5-kube-api-access-lxlzg\") pod \"nova-cell0-73a7-account-create-j6h4l\" (UID: \"f2ab3195-0fe3-46cf-b07b-ae48b41273e5\") " pod="openstack/nova-cell0-73a7-account-create-j6h4l" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.419213 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-73a7-account-create-j6h4l" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.449980 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2xgd\" (UniqueName: \"kubernetes.io/projected/26045870-04fb-4552-881d-eeb30715d747-kube-api-access-d2xgd\") pod \"nova-cell1-77da-account-create-wr7bc\" (UID: \"26045870-04fb-4552-881d-eeb30715d747\") " pod="openstack/nova-cell1-77da-account-create-wr7bc" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.552930 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2xgd\" (UniqueName: \"kubernetes.io/projected/26045870-04fb-4552-881d-eeb30715d747-kube-api-access-d2xgd\") pod \"nova-cell1-77da-account-create-wr7bc\" (UID: \"26045870-04fb-4552-881d-eeb30715d747\") " pod="openstack/nova-cell1-77da-account-create-wr7bc" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.572806 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2xgd\" (UniqueName: \"kubernetes.io/projected/26045870-04fb-4552-881d-eeb30715d747-kube-api-access-d2xgd\") pod \"nova-cell1-77da-account-create-wr7bc\" (UID: \"26045870-04fb-4552-881d-eeb30715d747\") " pod="openstack/nova-cell1-77da-account-create-wr7bc" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.637573 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-77da-account-create-wr7bc" Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.670771 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7720-account-create-78zkj"] Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.836991 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-73a7-account-create-j6h4l"] Oct 07 23:43:24 crc kubenswrapper[4871]: W1007 23:43:24.841949 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2ab3195_0fe3_46cf_b07b_ae48b41273e5.slice/crio-de1f511c4efb565dc0a47ccd23c62a2d98d87520f6e0dae5c111815dc9894785 WatchSource:0}: Error finding container de1f511c4efb565dc0a47ccd23c62a2d98d87520f6e0dae5c111815dc9894785: Status 404 returned error can't find the container with id de1f511c4efb565dc0a47ccd23c62a2d98d87520f6e0dae5c111815dc9894785 Oct 07 23:43:24 crc kubenswrapper[4871]: I1007 23:43:24.947723 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-77da-account-create-wr7bc"] Oct 07 23:43:25 crc kubenswrapper[4871]: W1007 23:43:25.002490 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26045870_04fb_4552_881d_eeb30715d747.slice/crio-8b01d41673005e667d2e2c7229f7becff2d2e66130140fc1636c874c73a8b22c WatchSource:0}: Error finding container 8b01d41673005e667d2e2c7229f7becff2d2e66130140fc1636c874c73a8b22c: Status 404 returned error can't find the container with id 8b01d41673005e667d2e2c7229f7becff2d2e66130140fc1636c874c73a8b22c Oct 07 23:43:25 crc kubenswrapper[4871]: I1007 23:43:25.040161 4871 generic.go:334] "Generic (PLEG): container finished" podID="833593af-531f-4b85-ab9f-56d6a657de34" containerID="9d2f7ea84742434bfbe6e84436cbd48d6a351ac96411bb9f8f7a5656404a1b7e" exitCode=0 Oct 07 23:43:25 crc kubenswrapper[4871]: I1007 23:43:25.040229 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7720-account-create-78zkj" event={"ID":"833593af-531f-4b85-ab9f-56d6a657de34","Type":"ContainerDied","Data":"9d2f7ea84742434bfbe6e84436cbd48d6a351ac96411bb9f8f7a5656404a1b7e"} Oct 07 23:43:25 crc kubenswrapper[4871]: I1007 23:43:25.040299 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7720-account-create-78zkj" event={"ID":"833593af-531f-4b85-ab9f-56d6a657de34","Type":"ContainerStarted","Data":"704d37d9446588f8eb7f44f4db1261d6636a9675b73ccfed3bd17070f211a336"} Oct 07 23:43:25 crc kubenswrapper[4871]: I1007 23:43:25.042057 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-73a7-account-create-j6h4l" event={"ID":"f2ab3195-0fe3-46cf-b07b-ae48b41273e5","Type":"ContainerStarted","Data":"de1f511c4efb565dc0a47ccd23c62a2d98d87520f6e0dae5c111815dc9894785"} Oct 07 23:43:25 crc kubenswrapper[4871]: I1007 23:43:25.043053 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-77da-account-create-wr7bc" event={"ID":"26045870-04fb-4552-881d-eeb30715d747","Type":"ContainerStarted","Data":"8b01d41673005e667d2e2c7229f7becff2d2e66130140fc1636c874c73a8b22c"} Oct 07 23:43:26 crc kubenswrapper[4871]: I1007 23:43:26.060362 4871 generic.go:334] "Generic (PLEG): container finished" podID="f2ab3195-0fe3-46cf-b07b-ae48b41273e5" containerID="3d127be49fc4e793064677b4ab32c550196edf416946fc832970b8a32dd8effd" exitCode=0 Oct 07 23:43:26 crc kubenswrapper[4871]: I1007 23:43:26.060709 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-73a7-account-create-j6h4l" event={"ID":"f2ab3195-0fe3-46cf-b07b-ae48b41273e5","Type":"ContainerDied","Data":"3d127be49fc4e793064677b4ab32c550196edf416946fc832970b8a32dd8effd"} Oct 07 23:43:26 crc kubenswrapper[4871]: I1007 23:43:26.062115 4871 generic.go:334] "Generic (PLEG): container finished" podID="26045870-04fb-4552-881d-eeb30715d747" containerID="47ae18d3fcf146c16f880e5af6617bf60cbfb09b064b3acc3aeb1ee071d589c7" exitCode=0 Oct 07 23:43:26 crc kubenswrapper[4871]: I1007 23:43:26.062377 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-77da-account-create-wr7bc" event={"ID":"26045870-04fb-4552-881d-eeb30715d747","Type":"ContainerDied","Data":"47ae18d3fcf146c16f880e5af6617bf60cbfb09b064b3acc3aeb1ee071d589c7"} Oct 07 23:43:26 crc kubenswrapper[4871]: I1007 23:43:26.397729 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7720-account-create-78zkj" Oct 07 23:43:26 crc kubenswrapper[4871]: I1007 23:43:26.498369 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz4fr\" (UniqueName: \"kubernetes.io/projected/833593af-531f-4b85-ab9f-56d6a657de34-kube-api-access-gz4fr\") pod \"833593af-531f-4b85-ab9f-56d6a657de34\" (UID: \"833593af-531f-4b85-ab9f-56d6a657de34\") " Oct 07 23:43:26 crc kubenswrapper[4871]: I1007 23:43:26.507507 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/833593af-531f-4b85-ab9f-56d6a657de34-kube-api-access-gz4fr" (OuterVolumeSpecName: "kube-api-access-gz4fr") pod "833593af-531f-4b85-ab9f-56d6a657de34" (UID: "833593af-531f-4b85-ab9f-56d6a657de34"). InnerVolumeSpecName "kube-api-access-gz4fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:26 crc kubenswrapper[4871]: I1007 23:43:26.600230 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz4fr\" (UniqueName: \"kubernetes.io/projected/833593af-531f-4b85-ab9f-56d6a657de34-kube-api-access-gz4fr\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.076874 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7720-account-create-78zkj" Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.076779 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7720-account-create-78zkj" event={"ID":"833593af-531f-4b85-ab9f-56d6a657de34","Type":"ContainerDied","Data":"704d37d9446588f8eb7f44f4db1261d6636a9675b73ccfed3bd17070f211a336"} Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.076981 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="704d37d9446588f8eb7f44f4db1261d6636a9675b73ccfed3bd17070f211a336" Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.550958 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-77da-account-create-wr7bc" Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.556846 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-73a7-account-create-j6h4l" Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.716730 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxlzg\" (UniqueName: \"kubernetes.io/projected/f2ab3195-0fe3-46cf-b07b-ae48b41273e5-kube-api-access-lxlzg\") pod \"f2ab3195-0fe3-46cf-b07b-ae48b41273e5\" (UID: \"f2ab3195-0fe3-46cf-b07b-ae48b41273e5\") " Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.716835 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2xgd\" (UniqueName: \"kubernetes.io/projected/26045870-04fb-4552-881d-eeb30715d747-kube-api-access-d2xgd\") pod \"26045870-04fb-4552-881d-eeb30715d747\" (UID: \"26045870-04fb-4552-881d-eeb30715d747\") " Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.725570 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26045870-04fb-4552-881d-eeb30715d747-kube-api-access-d2xgd" (OuterVolumeSpecName: "kube-api-access-d2xgd") pod "26045870-04fb-4552-881d-eeb30715d747" (UID: "26045870-04fb-4552-881d-eeb30715d747"). InnerVolumeSpecName "kube-api-access-d2xgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.731081 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2ab3195-0fe3-46cf-b07b-ae48b41273e5-kube-api-access-lxlzg" (OuterVolumeSpecName: "kube-api-access-lxlzg") pod "f2ab3195-0fe3-46cf-b07b-ae48b41273e5" (UID: "f2ab3195-0fe3-46cf-b07b-ae48b41273e5"). InnerVolumeSpecName "kube-api-access-lxlzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.819101 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2xgd\" (UniqueName: \"kubernetes.io/projected/26045870-04fb-4552-881d-eeb30715d747-kube-api-access-d2xgd\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:27 crc kubenswrapper[4871]: I1007 23:43:27.819148 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxlzg\" (UniqueName: \"kubernetes.io/projected/f2ab3195-0fe3-46cf-b07b-ae48b41273e5-kube-api-access-lxlzg\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:28 crc kubenswrapper[4871]: I1007 23:43:28.088691 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-73a7-account-create-j6h4l" event={"ID":"f2ab3195-0fe3-46cf-b07b-ae48b41273e5","Type":"ContainerDied","Data":"de1f511c4efb565dc0a47ccd23c62a2d98d87520f6e0dae5c111815dc9894785"} Oct 07 23:43:28 crc kubenswrapper[4871]: I1007 23:43:28.088733 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de1f511c4efb565dc0a47ccd23c62a2d98d87520f6e0dae5c111815dc9894785" Oct 07 23:43:28 crc kubenswrapper[4871]: I1007 23:43:28.088708 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-73a7-account-create-j6h4l" Oct 07 23:43:28 crc kubenswrapper[4871]: I1007 23:43:28.091837 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-77da-account-create-wr7bc" event={"ID":"26045870-04fb-4552-881d-eeb30715d747","Type":"ContainerDied","Data":"8b01d41673005e667d2e2c7229f7becff2d2e66130140fc1636c874c73a8b22c"} Oct 07 23:43:28 crc kubenswrapper[4871]: I1007 23:43:28.091877 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b01d41673005e667d2e2c7229f7becff2d2e66130140fc1636c874c73a8b22c" Oct 07 23:43:28 crc kubenswrapper[4871]: I1007 23:43:28.091920 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-77da-account-create-wr7bc" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.348868 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cn4jv"] Oct 07 23:43:29 crc kubenswrapper[4871]: E1007 23:43:29.349622 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26045870-04fb-4552-881d-eeb30715d747" containerName="mariadb-account-create" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.349638 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="26045870-04fb-4552-881d-eeb30715d747" containerName="mariadb-account-create" Oct 07 23:43:29 crc kubenswrapper[4871]: E1007 23:43:29.349678 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2ab3195-0fe3-46cf-b07b-ae48b41273e5" containerName="mariadb-account-create" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.349686 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2ab3195-0fe3-46cf-b07b-ae48b41273e5" containerName="mariadb-account-create" Oct 07 23:43:29 crc kubenswrapper[4871]: E1007 23:43:29.349703 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833593af-531f-4b85-ab9f-56d6a657de34" containerName="mariadb-account-create" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.349714 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="833593af-531f-4b85-ab9f-56d6a657de34" containerName="mariadb-account-create" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.349933 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2ab3195-0fe3-46cf-b07b-ae48b41273e5" containerName="mariadb-account-create" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.349955 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="26045870-04fb-4552-881d-eeb30715d747" containerName="mariadb-account-create" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.349969 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="833593af-531f-4b85-ab9f-56d6a657de34" containerName="mariadb-account-create" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.350677 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.353071 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.353106 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.353592 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qq6s8" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.365868 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cn4jv"] Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.549464 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8d24\" (UniqueName: \"kubernetes.io/projected/b06442f4-d46a-4ef2-a6d8-14855ec660c6-kube-api-access-m8d24\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.549672 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-scripts\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.549927 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-config-data\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.549966 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.651125 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-config-data\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.651179 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.651260 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8d24\" (UniqueName: \"kubernetes.io/projected/b06442f4-d46a-4ef2-a6d8-14855ec660c6-kube-api-access-m8d24\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.651321 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-scripts\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.656863 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-scripts\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.657758 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.658632 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-config-data\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.680347 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8d24\" (UniqueName: \"kubernetes.io/projected/b06442f4-d46a-4ef2-a6d8-14855ec660c6-kube-api-access-m8d24\") pod \"nova-cell0-conductor-db-sync-cn4jv\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:29 crc kubenswrapper[4871]: I1007 23:43:29.970577 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:30 crc kubenswrapper[4871]: I1007 23:43:30.454914 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cn4jv"] Oct 07 23:43:31 crc kubenswrapper[4871]: I1007 23:43:31.132843 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cn4jv" event={"ID":"b06442f4-d46a-4ef2-a6d8-14855ec660c6","Type":"ContainerStarted","Data":"35e0ab42533abbf5426d11d2001a272cd9a71eb7377859f377db8c5d1addeca9"} Oct 07 23:43:31 crc kubenswrapper[4871]: I1007 23:43:31.133180 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cn4jv" event={"ID":"b06442f4-d46a-4ef2-a6d8-14855ec660c6","Type":"ContainerStarted","Data":"f69cbf06cc12f85a0b428c3a117916c530f7b04c1f889b57481403007a345944"} Oct 07 23:43:31 crc kubenswrapper[4871]: I1007 23:43:31.158546 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-cn4jv" podStartSLOduration=2.158529595 podStartE2EDuration="2.158529595s" podCreationTimestamp="2025-10-07 23:43:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:31.152376302 +0000 UTC m=+5684.955074375" watchObservedRunningTime="2025-10-07 23:43:31.158529595 +0000 UTC m=+5684.961227668" Oct 07 23:43:35 crc kubenswrapper[4871]: I1007 23:43:35.513151 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:43:35 crc kubenswrapper[4871]: I1007 23:43:35.513998 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:43:35 crc kubenswrapper[4871]: I1007 23:43:35.514085 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:43:35 crc kubenswrapper[4871]: I1007 23:43:35.515425 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:43:35 crc kubenswrapper[4871]: I1007 23:43:35.515550 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" gracePeriod=600 Oct 07 23:43:35 crc kubenswrapper[4871]: E1007 23:43:35.640286 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:43:36 crc kubenswrapper[4871]: I1007 23:43:36.189224 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" exitCode=0 Oct 07 23:43:36 crc kubenswrapper[4871]: I1007 23:43:36.189316 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5"} Oct 07 23:43:36 crc kubenswrapper[4871]: I1007 23:43:36.190099 4871 scope.go:117] "RemoveContainer" containerID="a7d7ec4f805a57b5cbdcea4bbeab976af5520b960d9f0ff723e036600ebfeb03" Oct 07 23:43:36 crc kubenswrapper[4871]: I1007 23:43:36.191104 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:43:36 crc kubenswrapper[4871]: E1007 23:43:36.191651 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:43:36 crc kubenswrapper[4871]: I1007 23:43:36.192272 4871 generic.go:334] "Generic (PLEG): container finished" podID="b06442f4-d46a-4ef2-a6d8-14855ec660c6" containerID="35e0ab42533abbf5426d11d2001a272cd9a71eb7377859f377db8c5d1addeca9" exitCode=0 Oct 07 23:43:36 crc kubenswrapper[4871]: I1007 23:43:36.192363 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cn4jv" event={"ID":"b06442f4-d46a-4ef2-a6d8-14855ec660c6","Type":"ContainerDied","Data":"35e0ab42533abbf5426d11d2001a272cd9a71eb7377859f377db8c5d1addeca9"} Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.696421 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.720313 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8d24\" (UniqueName: \"kubernetes.io/projected/b06442f4-d46a-4ef2-a6d8-14855ec660c6-kube-api-access-m8d24\") pod \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.720474 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-config-data\") pod \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.720714 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-combined-ca-bundle\") pod \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.721910 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-scripts\") pod \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\" (UID: \"b06442f4-d46a-4ef2-a6d8-14855ec660c6\") " Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.728594 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-scripts" (OuterVolumeSpecName: "scripts") pod "b06442f4-d46a-4ef2-a6d8-14855ec660c6" (UID: "b06442f4-d46a-4ef2-a6d8-14855ec660c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.732088 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b06442f4-d46a-4ef2-a6d8-14855ec660c6-kube-api-access-m8d24" (OuterVolumeSpecName: "kube-api-access-m8d24") pod "b06442f4-d46a-4ef2-a6d8-14855ec660c6" (UID: "b06442f4-d46a-4ef2-a6d8-14855ec660c6"). InnerVolumeSpecName "kube-api-access-m8d24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.767065 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b06442f4-d46a-4ef2-a6d8-14855ec660c6" (UID: "b06442f4-d46a-4ef2-a6d8-14855ec660c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.774109 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-config-data" (OuterVolumeSpecName: "config-data") pod "b06442f4-d46a-4ef2-a6d8-14855ec660c6" (UID: "b06442f4-d46a-4ef2-a6d8-14855ec660c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.825405 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.825472 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.825493 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8d24\" (UniqueName: \"kubernetes.io/projected/b06442f4-d46a-4ef2-a6d8-14855ec660c6-kube-api-access-m8d24\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:37 crc kubenswrapper[4871]: I1007 23:43:37.825515 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b06442f4-d46a-4ef2-a6d8-14855ec660c6-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.247114 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cn4jv" event={"ID":"b06442f4-d46a-4ef2-a6d8-14855ec660c6","Type":"ContainerDied","Data":"f69cbf06cc12f85a0b428c3a117916c530f7b04c1f889b57481403007a345944"} Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.247205 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f69cbf06cc12f85a0b428c3a117916c530f7b04c1f889b57481403007a345944" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.247224 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cn4jv" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.362156 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 23:43:38 crc kubenswrapper[4871]: E1007 23:43:38.362936 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b06442f4-d46a-4ef2-a6d8-14855ec660c6" containerName="nova-cell0-conductor-db-sync" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.362977 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b06442f4-d46a-4ef2-a6d8-14855ec660c6" containerName="nova-cell0-conductor-db-sync" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.363482 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b06442f4-d46a-4ef2-a6d8-14855ec660c6" containerName="nova-cell0-conductor-db-sync" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.364659 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.367613 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qq6s8" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.370230 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.385668 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.437392 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.437465 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.437630 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljw74\" (UniqueName: \"kubernetes.io/projected/a69f3ff4-898a-4df6-9734-334249368868-kube-api-access-ljw74\") pod \"nova-cell0-conductor-0\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.539944 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljw74\" (UniqueName: \"kubernetes.io/projected/a69f3ff4-898a-4df6-9734-334249368868-kube-api-access-ljw74\") pod \"nova-cell0-conductor-0\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.540061 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.540104 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.551121 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.551649 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.562128 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljw74\" (UniqueName: \"kubernetes.io/projected/a69f3ff4-898a-4df6-9734-334249368868-kube-api-access-ljw74\") pod \"nova-cell0-conductor-0\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:38 crc kubenswrapper[4871]: I1007 23:43:38.683430 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:39 crc kubenswrapper[4871]: I1007 23:43:39.173743 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 23:43:39 crc kubenswrapper[4871]: W1007 23:43:39.177212 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda69f3ff4_898a_4df6_9734_334249368868.slice/crio-a7dfb752a70e6c420990ffe862c893a72e7c2b0016a74bab466cb62bf8a5d21b WatchSource:0}: Error finding container a7dfb752a70e6c420990ffe862c893a72e7c2b0016a74bab466cb62bf8a5d21b: Status 404 returned error can't find the container with id a7dfb752a70e6c420990ffe862c893a72e7c2b0016a74bab466cb62bf8a5d21b Oct 07 23:43:39 crc kubenswrapper[4871]: I1007 23:43:39.273234 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a69f3ff4-898a-4df6-9734-334249368868","Type":"ContainerStarted","Data":"a7dfb752a70e6c420990ffe862c893a72e7c2b0016a74bab466cb62bf8a5d21b"} Oct 07 23:43:40 crc kubenswrapper[4871]: I1007 23:43:40.289833 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a69f3ff4-898a-4df6-9734-334249368868","Type":"ContainerStarted","Data":"7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8"} Oct 07 23:43:40 crc kubenswrapper[4871]: I1007 23:43:40.290183 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:40 crc kubenswrapper[4871]: I1007 23:43:40.312379 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.312351619 podStartE2EDuration="2.312351619s" podCreationTimestamp="2025-10-07 23:43:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:40.311582999 +0000 UTC m=+5694.114281072" watchObservedRunningTime="2025-10-07 23:43:40.312351619 +0000 UTC m=+5694.115049722" Oct 07 23:43:48 crc kubenswrapper[4871]: I1007 23:43:48.730781 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.323601 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zpwpr"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.325225 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.327889 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.328146 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.341628 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zpwpr"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.377120 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.377260 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-scripts\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.377333 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc7cs\" (UniqueName: \"kubernetes.io/projected/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-kube-api-access-vc7cs\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.377419 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-config-data\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.457352 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.460755 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.463221 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.478666 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.478760 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-scripts\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.478882 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-config-data\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.478909 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc7cs\" (UniqueName: \"kubernetes.io/projected/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-kube-api-access-vc7cs\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.478933 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.478976 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/702a53e9-9112-4a87-8319-0147bf603fb9-logs\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.479007 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-config-data\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.479030 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpffv\" (UniqueName: \"kubernetes.io/projected/702a53e9-9112-4a87-8319-0147bf603fb9-kube-api-access-zpffv\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.487985 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-config-data\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.490756 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.494346 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.505211 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-scripts\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.511187 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc7cs\" (UniqueName: \"kubernetes.io/projected/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-kube-api-access-vc7cs\") pod \"nova-cell0-cell-mapping-zpwpr\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.547867 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.549348 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.555440 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.555995 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.557031 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.561469 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.567839 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.574647 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.597415 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.597537 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/702a53e9-9112-4a87-8319-0147bf603fb9-logs\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.597612 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhfgc\" (UniqueName: \"kubernetes.io/projected/bf928e42-2748-48f4-8044-194d804c19e5-kube-api-access-jhfgc\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.597655 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-config-data\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.597673 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.597691 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpffv\" (UniqueName: \"kubernetes.io/projected/702a53e9-9112-4a87-8319-0147bf603fb9-kube-api-access-zpffv\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.597849 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf928e42-2748-48f4-8044-194d804c19e5-logs\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.598813 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t77t9\" (UniqueName: \"kubernetes.io/projected/4aefc2ec-aa7f-4782-8993-f48aba3434d3-kube-api-access-t77t9\") pod \"nova-cell1-novncproxy-0\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.598954 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-config-data\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.599111 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.599141 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.599390 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/702a53e9-9112-4a87-8319-0147bf603fb9-logs\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.604682 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.610480 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-config-data\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.639569 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpffv\" (UniqueName: \"kubernetes.io/projected/702a53e9-9112-4a87-8319-0147bf603fb9-kube-api-access-zpffv\") pod \"nova-api-0\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.652834 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.658387 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.660104 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.662457 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.695117 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d6b768dc-jcjnc"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.696676 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.709334 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhfgc\" (UniqueName: \"kubernetes.io/projected/bf928e42-2748-48f4-8044-194d804c19e5-kube-api-access-jhfgc\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.709382 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.709417 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-config-data\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.709501 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf928e42-2748-48f4-8044-194d804c19e5-logs\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.709541 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t77t9\" (UniqueName: \"kubernetes.io/projected/4aefc2ec-aa7f-4782-8993-f48aba3434d3-kube-api-access-t77t9\") pod \"nova-cell1-novncproxy-0\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.709588 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.709605 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.716786 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf928e42-2748-48f4-8044-194d804c19e5-logs\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.719928 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.721036 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-config-data\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.721525 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.726379 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.733033 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhfgc\" (UniqueName: \"kubernetes.io/projected/bf928e42-2748-48f4-8044-194d804c19e5-kube-api-access-jhfgc\") pod \"nova-metadata-0\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.735051 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.738124 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d6b768dc-jcjnc"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.746373 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t77t9\" (UniqueName: \"kubernetes.io/projected/4aefc2ec-aa7f-4782-8993-f48aba3434d3-kube-api-access-t77t9\") pod \"nova-cell1-novncproxy-0\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.763017 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.773751 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.815083 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-config\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.817114 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-nb\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.817234 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpm5m\" (UniqueName: \"kubernetes.io/projected/be095c81-d042-480a-9000-f77699dcfdd6-kube-api-access-zpm5m\") pod \"nova-scheduler-0\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.818616 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-sb\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.818667 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-dns-svc\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.818735 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpjkv\" (UniqueName: \"kubernetes.io/projected/f2976020-e44a-44d7-890a-4dc6806d222a-kube-api-access-vpjkv\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.818769 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-config-data\") pod \"nova-scheduler-0\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.818856 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.892523 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.921565 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-config\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.921661 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-nb\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.921770 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpm5m\" (UniqueName: \"kubernetes.io/projected/be095c81-d042-480a-9000-f77699dcfdd6-kube-api-access-zpm5m\") pod \"nova-scheduler-0\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.921863 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-sb\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.921907 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-dns-svc\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.921952 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpjkv\" (UniqueName: \"kubernetes.io/projected/f2976020-e44a-44d7-890a-4dc6806d222a-kube-api-access-vpjkv\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.921986 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-config-data\") pod \"nova-scheduler-0\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.922027 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.924267 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-config\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.924300 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-dns-svc\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.924614 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-nb\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.925715 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-sb\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.929585 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.934261 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-config-data\") pod \"nova-scheduler-0\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.945686 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpm5m\" (UniqueName: \"kubernetes.io/projected/be095c81-d042-480a-9000-f77699dcfdd6-kube-api-access-zpm5m\") pod \"nova-scheduler-0\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " pod="openstack/nova-scheduler-0" Oct 07 23:43:49 crc kubenswrapper[4871]: I1007 23:43:49.947352 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpjkv\" (UniqueName: \"kubernetes.io/projected/f2976020-e44a-44d7-890a-4dc6806d222a-kube-api-access-vpjkv\") pod \"dnsmasq-dns-6d6b768dc-jcjnc\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.082991 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.089295 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.167363 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:43:50 crc kubenswrapper[4871]: W1007 23:43:50.169352 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf928e42_2748_48f4_8044_194d804c19e5.slice/crio-c3b33c2da6aa51bcc94944e29eb6747c9774f9f267bd354b61dbad3e7bc2baf7 WatchSource:0}: Error finding container c3b33c2da6aa51bcc94944e29eb6747c9774f9f267bd354b61dbad3e7bc2baf7: Status 404 returned error can't find the container with id c3b33c2da6aa51bcc94944e29eb6747c9774f9f267bd354b61dbad3e7bc2baf7 Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.309180 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zpwpr"] Oct 07 23:43:50 crc kubenswrapper[4871]: W1007 23:43:50.334305 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5e8b8e0_31d8_4c3a_a9e1_13f8f8d76d81.slice/crio-3757f3b3670324990d187f5dbd9c271bfcc9dc4a7e959546db617b68e0fc0357 WatchSource:0}: Error finding container 3757f3b3670324990d187f5dbd9c271bfcc9dc4a7e959546db617b68e0fc0357: Status 404 returned error can't find the container with id 3757f3b3670324990d187f5dbd9c271bfcc9dc4a7e959546db617b68e0fc0357 Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.388930 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.418564 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf928e42-2748-48f4-8044-194d804c19e5","Type":"ContainerStarted","Data":"fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff"} Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.418618 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf928e42-2748-48f4-8044-194d804c19e5","Type":"ContainerStarted","Data":"c3b33c2da6aa51bcc94944e29eb6747c9774f9f267bd354b61dbad3e7bc2baf7"} Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.419897 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zpwpr" event={"ID":"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81","Type":"ContainerStarted","Data":"3757f3b3670324990d187f5dbd9c271bfcc9dc4a7e959546db617b68e0fc0357"} Oct 07 23:43:50 crc kubenswrapper[4871]: W1007 23:43:50.430464 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aefc2ec_aa7f_4782_8993_f48aba3434d3.slice/crio-e1fdfbf3f8f0728600d38ba684bc42966ef206870eb0c114ce41964979d5ae19 WatchSource:0}: Error finding container e1fdfbf3f8f0728600d38ba684bc42966ef206870eb0c114ce41964979d5ae19: Status 404 returned error can't find the container with id e1fdfbf3f8f0728600d38ba684bc42966ef206870eb0c114ce41964979d5ae19 Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.514647 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:43:50 crc kubenswrapper[4871]: W1007 23:43:50.521760 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod702a53e9_9112_4a87_8319_0147bf603fb9.slice/crio-c2d3d628dec81a61dc050466217a8a437258f92521fe350f2f60bc174e27331c WatchSource:0}: Error finding container c2d3d628dec81a61dc050466217a8a437258f92521fe350f2f60bc174e27331c: Status 404 returned error can't find the container with id c2d3d628dec81a61dc050466217a8a437258f92521fe350f2f60bc174e27331c Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.662903 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdkfc"] Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.664036 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.667309 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.668839 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.669974 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdkfc"] Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.678225 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:43:50 crc kubenswrapper[4871]: W1007 23:43:50.699499 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe095c81_d042_480a_9000_f77699dcfdd6.slice/crio-fc9c93083c3b138ce1da50d314650a44a756f4fca0e8c34a2f128bae612546e3 WatchSource:0}: Error finding container fc9c93083c3b138ce1da50d314650a44a756f4fca0e8c34a2f128bae612546e3: Status 404 returned error can't find the container with id fc9c93083c3b138ce1da50d314650a44a756f4fca0e8c34a2f128bae612546e3 Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.768228 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d6b768dc-jcjnc"] Oct 07 23:43:50 crc kubenswrapper[4871]: W1007 23:43:50.777751 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2976020_e44a_44d7_890a_4dc6806d222a.slice/crio-fc2600253e49b6ccec2659681468c647502eede9639027041f005cb041e66e17 WatchSource:0}: Error finding container fc2600253e49b6ccec2659681468c647502eede9639027041f005cb041e66e17: Status 404 returned error can't find the container with id fc2600253e49b6ccec2659681468c647502eede9639027041f005cb041e66e17 Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.846704 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-config-data\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.847324 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-scripts\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.847524 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6dt9\" (UniqueName: \"kubernetes.io/projected/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-kube-api-access-p6dt9\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.847631 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.949904 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6dt9\" (UniqueName: \"kubernetes.io/projected/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-kube-api-access-p6dt9\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.949963 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.950012 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-config-data\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.950031 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-scripts\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.956567 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.961117 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-scripts\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.962415 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-config-data\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.970612 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6dt9\" (UniqueName: \"kubernetes.io/projected/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-kube-api-access-p6dt9\") pod \"nova-cell1-conductor-db-sync-xdkfc\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:50 crc kubenswrapper[4871]: I1007 23:43:50.983468 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:43:50 crc kubenswrapper[4871]: E1007 23:43:50.983684 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.182878 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.463213 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf928e42-2748-48f4-8044-194d804c19e5","Type":"ContainerStarted","Data":"cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.474409 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be095c81-d042-480a-9000-f77699dcfdd6","Type":"ContainerStarted","Data":"c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.474454 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be095c81-d042-480a-9000-f77699dcfdd6","Type":"ContainerStarted","Data":"fc9c93083c3b138ce1da50d314650a44a756f4fca0e8c34a2f128bae612546e3"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.483445 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zpwpr" event={"ID":"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81","Type":"ContainerStarted","Data":"8d1bf31d634b04a296616b1251e9785c15f4a8d5b038e27a0dd45a18ba86a556"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.486555 4871 generic.go:334] "Generic (PLEG): container finished" podID="f2976020-e44a-44d7-890a-4dc6806d222a" containerID="7e654d77532b9ec11568b64fb8c2d71d4685ac1b9b4d4a1d8794ec17fc62f32b" exitCode=0 Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.486641 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" event={"ID":"f2976020-e44a-44d7-890a-4dc6806d222a","Type":"ContainerDied","Data":"7e654d77532b9ec11568b64fb8c2d71d4685ac1b9b4d4a1d8794ec17fc62f32b"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.486661 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" event={"ID":"f2976020-e44a-44d7-890a-4dc6806d222a","Type":"ContainerStarted","Data":"fc2600253e49b6ccec2659681468c647502eede9639027041f005cb041e66e17"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.489620 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"702a53e9-9112-4a87-8319-0147bf603fb9","Type":"ContainerStarted","Data":"c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.489648 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"702a53e9-9112-4a87-8319-0147bf603fb9","Type":"ContainerStarted","Data":"c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.489658 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"702a53e9-9112-4a87-8319-0147bf603fb9","Type":"ContainerStarted","Data":"c2d3d628dec81a61dc050466217a8a437258f92521fe350f2f60bc174e27331c"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.489861 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.489840559 podStartE2EDuration="2.489840559s" podCreationTimestamp="2025-10-07 23:43:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:51.482239108 +0000 UTC m=+5705.284937181" watchObservedRunningTime="2025-10-07 23:43:51.489840559 +0000 UTC m=+5705.292538622" Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.498994 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.49897646 podStartE2EDuration="2.49897646s" podCreationTimestamp="2025-10-07 23:43:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:51.498685432 +0000 UTC m=+5705.301383505" watchObservedRunningTime="2025-10-07 23:43:51.49897646 +0000 UTC m=+5705.301674533" Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.499035 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4aefc2ec-aa7f-4782-8993-f48aba3434d3","Type":"ContainerStarted","Data":"5af235543317d416b3ea0b9fb8edceb8e1d2aca6a85e1c4f511e1521b38236f9"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.499071 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4aefc2ec-aa7f-4782-8993-f48aba3434d3","Type":"ContainerStarted","Data":"e1fdfbf3f8f0728600d38ba684bc42966ef206870eb0c114ce41964979d5ae19"} Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.536923 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.536904951 podStartE2EDuration="2.536904951s" podCreationTimestamp="2025-10-07 23:43:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:51.527811731 +0000 UTC m=+5705.330509794" watchObservedRunningTime="2025-10-07 23:43:51.536904951 +0000 UTC m=+5705.339603024" Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.548122 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zpwpr" podStartSLOduration=2.548102977 podStartE2EDuration="2.548102977s" podCreationTimestamp="2025-10-07 23:43:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:51.543267279 +0000 UTC m=+5705.345965352" watchObservedRunningTime="2025-10-07 23:43:51.548102977 +0000 UTC m=+5705.350801050" Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.646134 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.646117304 podStartE2EDuration="2.646117304s" podCreationTimestamp="2025-10-07 23:43:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:51.565913417 +0000 UTC m=+5705.368611490" watchObservedRunningTime="2025-10-07 23:43:51.646117304 +0000 UTC m=+5705.448815377" Oct 07 23:43:51 crc kubenswrapper[4871]: I1007 23:43:51.649135 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdkfc"] Oct 07 23:43:52 crc kubenswrapper[4871]: I1007 23:43:52.515346 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xdkfc" event={"ID":"ae15ef8e-e5a8-48ae-9317-5b0bbd197508","Type":"ContainerStarted","Data":"0c1410f98ef3030ed6237b7719b3e4aaaed7b574b2c55523e5d535c73c9c67ab"} Oct 07 23:43:52 crc kubenswrapper[4871]: I1007 23:43:52.515768 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xdkfc" event={"ID":"ae15ef8e-e5a8-48ae-9317-5b0bbd197508","Type":"ContainerStarted","Data":"3f21d16bdb18b8b8b4fbbdd83cba1c924f45c7993273f4012dbdcfb0cdebe15c"} Oct 07 23:43:52 crc kubenswrapper[4871]: I1007 23:43:52.519431 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" event={"ID":"f2976020-e44a-44d7-890a-4dc6806d222a","Type":"ContainerStarted","Data":"e5bb5fed9ae22642b24d7397e58b83db9bd31e81f93799c2e1551bb3ff7d34ef"} Oct 07 23:43:52 crc kubenswrapper[4871]: I1007 23:43:52.540324 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-xdkfc" podStartSLOduration=2.540300079 podStartE2EDuration="2.540300079s" podCreationTimestamp="2025-10-07 23:43:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:52.535486502 +0000 UTC m=+5706.338184615" watchObservedRunningTime="2025-10-07 23:43:52.540300079 +0000 UTC m=+5706.342998192" Oct 07 23:43:52 crc kubenswrapper[4871]: I1007 23:43:52.575937 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" podStartSLOduration=3.575908139 podStartE2EDuration="3.575908139s" podCreationTimestamp="2025-10-07 23:43:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:52.567181459 +0000 UTC m=+5706.369879532" watchObservedRunningTime="2025-10-07 23:43:52.575908139 +0000 UTC m=+5706.378606242" Oct 07 23:43:53 crc kubenswrapper[4871]: I1007 23:43:53.532501 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:43:54 crc kubenswrapper[4871]: I1007 23:43:54.545360 4871 generic.go:334] "Generic (PLEG): container finished" podID="ae15ef8e-e5a8-48ae-9317-5b0bbd197508" containerID="0c1410f98ef3030ed6237b7719b3e4aaaed7b574b2c55523e5d535c73c9c67ab" exitCode=0 Oct 07 23:43:54 crc kubenswrapper[4871]: I1007 23:43:54.545470 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xdkfc" event={"ID":"ae15ef8e-e5a8-48ae-9317-5b0bbd197508","Type":"ContainerDied","Data":"0c1410f98ef3030ed6237b7719b3e4aaaed7b574b2c55523e5d535c73c9c67ab"} Oct 07 23:43:54 crc kubenswrapper[4871]: I1007 23:43:54.735606 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 23:43:54 crc kubenswrapper[4871]: I1007 23:43:54.735850 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 23:43:54 crc kubenswrapper[4871]: I1007 23:43:54.764005 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.083580 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.290659 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-644z4"] Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.293024 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.300363 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-644z4"] Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.439945 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgr66\" (UniqueName: \"kubernetes.io/projected/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-kube-api-access-fgr66\") pod \"certified-operators-644z4\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.440497 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-utilities\") pod \"certified-operators-644z4\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.440548 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-catalog-content\") pod \"certified-operators-644z4\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.543070 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgr66\" (UniqueName: \"kubernetes.io/projected/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-kube-api-access-fgr66\") pod \"certified-operators-644z4\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.543354 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-utilities\") pod \"certified-operators-644z4\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.543394 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-catalog-content\") pod \"certified-operators-644z4\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.544156 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-utilities\") pod \"certified-operators-644z4\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.544284 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-catalog-content\") pod \"certified-operators-644z4\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.559008 4871 generic.go:334] "Generic (PLEG): container finished" podID="e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81" containerID="8d1bf31d634b04a296616b1251e9785c15f4a8d5b038e27a0dd45a18ba86a556" exitCode=0 Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.559084 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zpwpr" event={"ID":"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81","Type":"ContainerDied","Data":"8d1bf31d634b04a296616b1251e9785c15f4a8d5b038e27a0dd45a18ba86a556"} Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.573464 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgr66\" (UniqueName: \"kubernetes.io/projected/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-kube-api-access-fgr66\") pod \"certified-operators-644z4\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.618597 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:43:55 crc kubenswrapper[4871]: I1007 23:43:55.995390 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.159858 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-config-data\") pod \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.159955 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-combined-ca-bundle\") pod \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.160001 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-scripts\") pod \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.160098 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6dt9\" (UniqueName: \"kubernetes.io/projected/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-kube-api-access-p6dt9\") pod \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\" (UID: \"ae15ef8e-e5a8-48ae-9317-5b0bbd197508\") " Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.165147 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-scripts" (OuterVolumeSpecName: "scripts") pod "ae15ef8e-e5a8-48ae-9317-5b0bbd197508" (UID: "ae15ef8e-e5a8-48ae-9317-5b0bbd197508"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.165181 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-kube-api-access-p6dt9" (OuterVolumeSpecName: "kube-api-access-p6dt9") pod "ae15ef8e-e5a8-48ae-9317-5b0bbd197508" (UID: "ae15ef8e-e5a8-48ae-9317-5b0bbd197508"). InnerVolumeSpecName "kube-api-access-p6dt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.182722 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-config-data" (OuterVolumeSpecName: "config-data") pod "ae15ef8e-e5a8-48ae-9317-5b0bbd197508" (UID: "ae15ef8e-e5a8-48ae-9317-5b0bbd197508"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.201736 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae15ef8e-e5a8-48ae-9317-5b0bbd197508" (UID: "ae15ef8e-e5a8-48ae-9317-5b0bbd197508"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.214287 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-644z4"] Oct 07 23:43:56 crc kubenswrapper[4871]: W1007 23:43:56.218072 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a6ef7d8_24dd_4edb_b492_b430adf2b67f.slice/crio-d4cbef210d33012659a8d50751e2ab4f604426e6888f33eb62959e3c774b45ed WatchSource:0}: Error finding container d4cbef210d33012659a8d50751e2ab4f604426e6888f33eb62959e3c774b45ed: Status 404 returned error can't find the container with id d4cbef210d33012659a8d50751e2ab4f604426e6888f33eb62959e3c774b45ed Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.263886 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.263929 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.263943 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.263956 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6dt9\" (UniqueName: \"kubernetes.io/projected/ae15ef8e-e5a8-48ae-9317-5b0bbd197508-kube-api-access-p6dt9\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.591329 4871 generic.go:334] "Generic (PLEG): container finished" podID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerID="1df1ebc05b4ef76062421e0f9bd10566b96ca314d967847975c13682218398c0" exitCode=0 Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.591929 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-644z4" event={"ID":"8a6ef7d8-24dd-4edb-b492-b430adf2b67f","Type":"ContainerDied","Data":"1df1ebc05b4ef76062421e0f9bd10566b96ca314d967847975c13682218398c0"} Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.591977 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-644z4" event={"ID":"8a6ef7d8-24dd-4edb-b492-b430adf2b67f","Type":"ContainerStarted","Data":"d4cbef210d33012659a8d50751e2ab4f604426e6888f33eb62959e3c774b45ed"} Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.600209 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.604651 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xdkfc" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.604843 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xdkfc" event={"ID":"ae15ef8e-e5a8-48ae-9317-5b0bbd197508","Type":"ContainerDied","Data":"3f21d16bdb18b8b8b4fbbdd83cba1c924f45c7993273f4012dbdcfb0cdebe15c"} Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.604879 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f21d16bdb18b8b8b4fbbdd83cba1c924f45c7993273f4012dbdcfb0cdebe15c" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.673994 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 23:43:56 crc kubenswrapper[4871]: E1007 23:43:56.674495 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae15ef8e-e5a8-48ae-9317-5b0bbd197508" containerName="nova-cell1-conductor-db-sync" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.674513 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae15ef8e-e5a8-48ae-9317-5b0bbd197508" containerName="nova-cell1-conductor-db-sync" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.674743 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae15ef8e-e5a8-48ae-9317-5b0bbd197508" containerName="nova-cell1-conductor-db-sync" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.675602 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.678859 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.687648 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.802607 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.802898 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.803038 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdffk\" (UniqueName: \"kubernetes.io/projected/88b2fbdc-ee2b-4803-be96-c47713900f08-kube-api-access-tdffk\") pod \"nova-cell1-conductor-0\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.904463 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.904599 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.904646 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdffk\" (UniqueName: \"kubernetes.io/projected/88b2fbdc-ee2b-4803-be96-c47713900f08-kube-api-access-tdffk\") pod \"nova-cell1-conductor-0\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.911239 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.917219 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.921096 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdffk\" (UniqueName: \"kubernetes.io/projected/88b2fbdc-ee2b-4803-be96-c47713900f08-kube-api-access-tdffk\") pod \"nova-cell1-conductor-0\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:56 crc kubenswrapper[4871]: I1007 23:43:56.995602 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.001697 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.107845 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vc7cs\" (UniqueName: \"kubernetes.io/projected/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-kube-api-access-vc7cs\") pod \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.107974 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-scripts\") pod \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.108045 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-combined-ca-bundle\") pod \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.108091 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-config-data\") pod \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\" (UID: \"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81\") " Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.114298 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-scripts" (OuterVolumeSpecName: "scripts") pod "e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81" (UID: "e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.114311 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-kube-api-access-vc7cs" (OuterVolumeSpecName: "kube-api-access-vc7cs") pod "e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81" (UID: "e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81"). InnerVolumeSpecName "kube-api-access-vc7cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.133885 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81" (UID: "e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.139593 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-config-data" (OuterVolumeSpecName: "config-data") pod "e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81" (UID: "e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.211186 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vc7cs\" (UniqueName: \"kubernetes.io/projected/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-kube-api-access-vc7cs\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.211234 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.211255 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.211269 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.447557 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 23:43:57 crc kubenswrapper[4871]: W1007 23:43:57.455321 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88b2fbdc_ee2b_4803_be96_c47713900f08.slice/crio-d7b7d7bfe6b672bc79a7bd2efa33caa19d2ae7374b942d25b938feda4ad01e6e WatchSource:0}: Error finding container d7b7d7bfe6b672bc79a7bd2efa33caa19d2ae7374b942d25b938feda4ad01e6e: Status 404 returned error can't find the container with id d7b7d7bfe6b672bc79a7bd2efa33caa19d2ae7374b942d25b938feda4ad01e6e Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.627565 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"88b2fbdc-ee2b-4803-be96-c47713900f08","Type":"ContainerStarted","Data":"d7b7d7bfe6b672bc79a7bd2efa33caa19d2ae7374b942d25b938feda4ad01e6e"} Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.631278 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zpwpr" event={"ID":"e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81","Type":"ContainerDied","Data":"3757f3b3670324990d187f5dbd9c271bfcc9dc4a7e959546db617b68e0fc0357"} Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.631336 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zpwpr" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.631343 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3757f3b3670324990d187f5dbd9c271bfcc9dc4a7e959546db617b68e0fc0357" Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.761385 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.761686 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="702a53e9-9112-4a87-8319-0147bf603fb9" containerName="nova-api-log" containerID="cri-o://c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff" gracePeriod=30 Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.761957 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="702a53e9-9112-4a87-8319-0147bf603fb9" containerName="nova-api-api" containerID="cri-o://c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564" gracePeriod=30 Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.770671 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.770926 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="be095c81-d042-480a-9000-f77699dcfdd6" containerName="nova-scheduler-scheduler" containerID="cri-o://c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc" gracePeriod=30 Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.786687 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.786987 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bf928e42-2748-48f4-8044-194d804c19e5" containerName="nova-metadata-log" containerID="cri-o://fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff" gracePeriod=30 Oct 07 23:43:57 crc kubenswrapper[4871]: I1007 23:43:57.787498 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bf928e42-2748-48f4-8044-194d804c19e5" containerName="nova-metadata-metadata" containerID="cri-o://cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465" gracePeriod=30 Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.296189 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.322500 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439010 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-config-data\") pod \"bf928e42-2748-48f4-8044-194d804c19e5\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439075 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-config-data\") pod \"702a53e9-9112-4a87-8319-0147bf603fb9\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439117 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpffv\" (UniqueName: \"kubernetes.io/projected/702a53e9-9112-4a87-8319-0147bf603fb9-kube-api-access-zpffv\") pod \"702a53e9-9112-4a87-8319-0147bf603fb9\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439198 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf928e42-2748-48f4-8044-194d804c19e5-logs\") pod \"bf928e42-2748-48f4-8044-194d804c19e5\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439219 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/702a53e9-9112-4a87-8319-0147bf603fb9-logs\") pod \"702a53e9-9112-4a87-8319-0147bf603fb9\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439292 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-combined-ca-bundle\") pod \"bf928e42-2748-48f4-8044-194d804c19e5\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439321 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhfgc\" (UniqueName: \"kubernetes.io/projected/bf928e42-2748-48f4-8044-194d804c19e5-kube-api-access-jhfgc\") pod \"bf928e42-2748-48f4-8044-194d804c19e5\" (UID: \"bf928e42-2748-48f4-8044-194d804c19e5\") " Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439351 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-combined-ca-bundle\") pod \"702a53e9-9112-4a87-8319-0147bf603fb9\" (UID: \"702a53e9-9112-4a87-8319-0147bf603fb9\") " Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439750 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/702a53e9-9112-4a87-8319-0147bf603fb9-logs" (OuterVolumeSpecName: "logs") pod "702a53e9-9112-4a87-8319-0147bf603fb9" (UID: "702a53e9-9112-4a87-8319-0147bf603fb9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.439854 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf928e42-2748-48f4-8044-194d804c19e5-logs" (OuterVolumeSpecName: "logs") pod "bf928e42-2748-48f4-8044-194d804c19e5" (UID: "bf928e42-2748-48f4-8044-194d804c19e5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.445063 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702a53e9-9112-4a87-8319-0147bf603fb9-kube-api-access-zpffv" (OuterVolumeSpecName: "kube-api-access-zpffv") pod "702a53e9-9112-4a87-8319-0147bf603fb9" (UID: "702a53e9-9112-4a87-8319-0147bf603fb9"). InnerVolumeSpecName "kube-api-access-zpffv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.446140 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf928e42-2748-48f4-8044-194d804c19e5-kube-api-access-jhfgc" (OuterVolumeSpecName: "kube-api-access-jhfgc") pod "bf928e42-2748-48f4-8044-194d804c19e5" (UID: "bf928e42-2748-48f4-8044-194d804c19e5"). InnerVolumeSpecName "kube-api-access-jhfgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.464673 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf928e42-2748-48f4-8044-194d804c19e5" (UID: "bf928e42-2748-48f4-8044-194d804c19e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.465081 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "702a53e9-9112-4a87-8319-0147bf603fb9" (UID: "702a53e9-9112-4a87-8319-0147bf603fb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.473394 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-config-data" (OuterVolumeSpecName: "config-data") pod "702a53e9-9112-4a87-8319-0147bf603fb9" (UID: "702a53e9-9112-4a87-8319-0147bf603fb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.476505 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-config-data" (OuterVolumeSpecName: "config-data") pod "bf928e42-2748-48f4-8044-194d804c19e5" (UID: "bf928e42-2748-48f4-8044-194d804c19e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.541264 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.541294 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.541305 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpffv\" (UniqueName: \"kubernetes.io/projected/702a53e9-9112-4a87-8319-0147bf603fb9-kube-api-access-zpffv\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.541314 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf928e42-2748-48f4-8044-194d804c19e5-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.541321 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/702a53e9-9112-4a87-8319-0147bf603fb9-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.541329 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf928e42-2748-48f4-8044-194d804c19e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.541336 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhfgc\" (UniqueName: \"kubernetes.io/projected/bf928e42-2748-48f4-8044-194d804c19e5-kube-api-access-jhfgc\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.541344 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702a53e9-9112-4a87-8319-0147bf603fb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.661340 4871 generic.go:334] "Generic (PLEG): container finished" podID="bf928e42-2748-48f4-8044-194d804c19e5" containerID="cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465" exitCode=0 Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.661913 4871 generic.go:334] "Generic (PLEG): container finished" podID="bf928e42-2748-48f4-8044-194d804c19e5" containerID="fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff" exitCode=143 Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.661404 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf928e42-2748-48f4-8044-194d804c19e5","Type":"ContainerDied","Data":"cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465"} Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.661490 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.661980 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf928e42-2748-48f4-8044-194d804c19e5","Type":"ContainerDied","Data":"fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff"} Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.662006 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf928e42-2748-48f4-8044-194d804c19e5","Type":"ContainerDied","Data":"c3b33c2da6aa51bcc94944e29eb6747c9774f9f267bd354b61dbad3e7bc2baf7"} Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.662010 4871 scope.go:117] "RemoveContainer" containerID="cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.666942 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"88b2fbdc-ee2b-4803-be96-c47713900f08","Type":"ContainerStarted","Data":"b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a"} Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.667051 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.671099 4871 generic.go:334] "Generic (PLEG): container finished" podID="702a53e9-9112-4a87-8319-0147bf603fb9" containerID="c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564" exitCode=0 Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.671165 4871 generic.go:334] "Generic (PLEG): container finished" podID="702a53e9-9112-4a87-8319-0147bf603fb9" containerID="c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff" exitCode=143 Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.671129 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"702a53e9-9112-4a87-8319-0147bf603fb9","Type":"ContainerDied","Data":"c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564"} Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.671222 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.671233 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"702a53e9-9112-4a87-8319-0147bf603fb9","Type":"ContainerDied","Data":"c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff"} Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.671359 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"702a53e9-9112-4a87-8319-0147bf603fb9","Type":"ContainerDied","Data":"c2d3d628dec81a61dc050466217a8a437258f92521fe350f2f60bc174e27331c"} Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.698879 4871 scope.go:117] "RemoveContainer" containerID="fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.708812 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.708778733 podStartE2EDuration="2.708778733s" podCreationTimestamp="2025-10-07 23:43:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:43:58.690306325 +0000 UTC m=+5712.493004418" watchObservedRunningTime="2025-10-07 23:43:58.708778733 +0000 UTC m=+5712.511476806" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.715947 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.733488 4871 scope.go:117] "RemoveContainer" containerID="cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.733620 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:43:58 crc kubenswrapper[4871]: E1007 23:43:58.734122 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465\": container with ID starting with cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465 not found: ID does not exist" containerID="cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.734163 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465"} err="failed to get container status \"cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465\": rpc error: code = NotFound desc = could not find container \"cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465\": container with ID starting with cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465 not found: ID does not exist" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.734190 4871 scope.go:117] "RemoveContainer" containerID="fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff" Oct 07 23:43:58 crc kubenswrapper[4871]: E1007 23:43:58.734667 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff\": container with ID starting with fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff not found: ID does not exist" containerID="fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.734696 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff"} err="failed to get container status \"fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff\": rpc error: code = NotFound desc = could not find container \"fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff\": container with ID starting with fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff not found: ID does not exist" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.734714 4871 scope.go:117] "RemoveContainer" containerID="cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.735034 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465"} err="failed to get container status \"cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465\": rpc error: code = NotFound desc = could not find container \"cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465\": container with ID starting with cd8177976f471fc8a6269dc2825211df2a9151a88ecd458c00a1b0626effe465 not found: ID does not exist" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.735075 4871 scope.go:117] "RemoveContainer" containerID="fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.735337 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff"} err="failed to get container status \"fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff\": rpc error: code = NotFound desc = could not find container \"fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff\": container with ID starting with fbfdd595a69a6586ba61a7d6e7e8e68db531a7f2e567415101aeb813028794ff not found: ID does not exist" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.735364 4871 scope.go:117] "RemoveContainer" containerID="c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.751864 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.768777 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 23:43:58 crc kubenswrapper[4871]: E1007 23:43:58.769397 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702a53e9-9112-4a87-8319-0147bf603fb9" containerName="nova-api-api" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769420 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="702a53e9-9112-4a87-8319-0147bf603fb9" containerName="nova-api-api" Oct 07 23:43:58 crc kubenswrapper[4871]: E1007 23:43:58.769438 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702a53e9-9112-4a87-8319-0147bf603fb9" containerName="nova-api-log" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769446 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="702a53e9-9112-4a87-8319-0147bf603fb9" containerName="nova-api-log" Oct 07 23:43:58 crc kubenswrapper[4871]: E1007 23:43:58.769460 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81" containerName="nova-manage" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769468 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81" containerName="nova-manage" Oct 07 23:43:58 crc kubenswrapper[4871]: E1007 23:43:58.769493 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf928e42-2748-48f4-8044-194d804c19e5" containerName="nova-metadata-log" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769502 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf928e42-2748-48f4-8044-194d804c19e5" containerName="nova-metadata-log" Oct 07 23:43:58 crc kubenswrapper[4871]: E1007 23:43:58.769513 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf928e42-2748-48f4-8044-194d804c19e5" containerName="nova-metadata-metadata" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769521 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf928e42-2748-48f4-8044-194d804c19e5" containerName="nova-metadata-metadata" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769750 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="702a53e9-9112-4a87-8319-0147bf603fb9" containerName="nova-api-api" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769777 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="702a53e9-9112-4a87-8319-0147bf603fb9" containerName="nova-api-log" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769807 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf928e42-2748-48f4-8044-194d804c19e5" containerName="nova-metadata-log" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769822 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81" containerName="nova-manage" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.769836 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf928e42-2748-48f4-8044-194d804c19e5" containerName="nova-metadata-metadata" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.771050 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.783501 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.785668 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.795171 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.799469 4871 scope.go:117] "RemoveContainer" containerID="c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.804454 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.806347 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.811259 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.824170 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.848648 4871 scope.go:117] "RemoveContainer" containerID="c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564" Oct 07 23:43:58 crc kubenswrapper[4871]: E1007 23:43:58.849452 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564\": container with ID starting with c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564 not found: ID does not exist" containerID="c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.849485 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564"} err="failed to get container status \"c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564\": rpc error: code = NotFound desc = could not find container \"c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564\": container with ID starting with c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564 not found: ID does not exist" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.849509 4871 scope.go:117] "RemoveContainer" containerID="c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff" Oct 07 23:43:58 crc kubenswrapper[4871]: E1007 23:43:58.849986 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff\": container with ID starting with c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff not found: ID does not exist" containerID="c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.850023 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff"} err="failed to get container status \"c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff\": rpc error: code = NotFound desc = could not find container \"c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff\": container with ID starting with c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff not found: ID does not exist" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.850043 4871 scope.go:117] "RemoveContainer" containerID="c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.850900 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564"} err="failed to get container status \"c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564\": rpc error: code = NotFound desc = could not find container \"c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564\": container with ID starting with c5f5b70d189767068f2395cfd4c30a4e73c77cc2c51caeeffababb29951ba564 not found: ID does not exist" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.850926 4871 scope.go:117] "RemoveContainer" containerID="c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.851146 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff"} err="failed to get container status \"c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff\": rpc error: code = NotFound desc = could not find container \"c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff\": container with ID starting with c1a3e469d15b14ffdee1712ccf3a0f728a9ca1395364b3ad293c8dbd5008b2ff not found: ID does not exist" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.951572 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-config-data\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.951669 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-config-data\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.951697 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f72129-cdf2-48c2-92e1-555817d6e25b-logs\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.951711 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.951761 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f4kh\" (UniqueName: \"kubernetes.io/projected/76f72129-cdf2-48c2-92e1-555817d6e25b-kube-api-access-5f4kh\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.951780 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhq55\" (UniqueName: \"kubernetes.io/projected/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-kube-api-access-jhq55\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.951872 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-logs\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.951900 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.992976 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="702a53e9-9112-4a87-8319-0147bf603fb9" path="/var/lib/kubelet/pods/702a53e9-9112-4a87-8319-0147bf603fb9/volumes" Oct 07 23:43:58 crc kubenswrapper[4871]: I1007 23:43:58.993589 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf928e42-2748-48f4-8044-194d804c19e5" path="/var/lib/kubelet/pods/bf928e42-2748-48f4-8044-194d804c19e5/volumes" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.053333 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-config-data\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.053465 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-config-data\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.053500 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f72129-cdf2-48c2-92e1-555817d6e25b-logs\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.053521 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.053589 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f4kh\" (UniqueName: \"kubernetes.io/projected/76f72129-cdf2-48c2-92e1-555817d6e25b-kube-api-access-5f4kh\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.053613 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhq55\" (UniqueName: \"kubernetes.io/projected/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-kube-api-access-jhq55\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.053636 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-logs\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.053669 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.053970 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f72129-cdf2-48c2-92e1-555817d6e25b-logs\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.054156 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-logs\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.058117 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-config-data\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.058263 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-config-data\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.059983 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.066873 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.069499 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhq55\" (UniqueName: \"kubernetes.io/projected/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-kube-api-access-jhq55\") pod \"nova-api-0\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " pod="openstack/nova-api-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.069760 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f4kh\" (UniqueName: \"kubernetes.io/projected/76f72129-cdf2-48c2-92e1-555817d6e25b-kube-api-access-5f4kh\") pod \"nova-metadata-0\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " pod="openstack/nova-metadata-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.098711 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.145281 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.614507 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.662855 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:43:59 crc kubenswrapper[4871]: W1007 23:43:59.671350 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76f72129_cdf2_48c2_92e1_555817d6e25b.slice/crio-28ca81145da334b1f25dce453221b2f29524112eb0eddb56ca708106f254a884 WatchSource:0}: Error finding container 28ca81145da334b1f25dce453221b2f29524112eb0eddb56ca708106f254a884: Status 404 returned error can't find the container with id 28ca81145da334b1f25dce453221b2f29524112eb0eddb56ca708106f254a884 Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.683408 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6","Type":"ContainerStarted","Data":"fa30f725c0b73dac68fd53bdc17389b383d71b7cd4312a25a7dab36ad0fb5222"} Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.763822 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:43:59 crc kubenswrapper[4871]: I1007 23:43:59.779932 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:44:00 crc kubenswrapper[4871]: I1007 23:44:00.091695 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:44:00 crc kubenswrapper[4871]: I1007 23:44:00.155651 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b67b947d7-cks8k"] Oct 07 23:44:00 crc kubenswrapper[4871]: I1007 23:44:00.156202 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" podUID="00e42cab-5504-46f1-a7ad-a74f4007bfef" containerName="dnsmasq-dns" containerID="cri-o://8885cb7238f5b3a02820c42f2a0544f7b5ff9f92b5de2a2fab0f68d83810b600" gracePeriod=10 Oct 07 23:44:00 crc kubenswrapper[4871]: I1007 23:44:00.700831 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76f72129-cdf2-48c2-92e1-555817d6e25b","Type":"ContainerStarted","Data":"042233c1ac1f908b0a89151b3a0863aa28ba0813996b31e2fa2e7dc378a11212"} Oct 07 23:44:00 crc kubenswrapper[4871]: I1007 23:44:00.700883 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76f72129-cdf2-48c2-92e1-555817d6e25b","Type":"ContainerStarted","Data":"28ca81145da334b1f25dce453221b2f29524112eb0eddb56ca708106f254a884"} Oct 07 23:44:00 crc kubenswrapper[4871]: I1007 23:44:00.702254 4871 generic.go:334] "Generic (PLEG): container finished" podID="00e42cab-5504-46f1-a7ad-a74f4007bfef" containerID="8885cb7238f5b3a02820c42f2a0544f7b5ff9f92b5de2a2fab0f68d83810b600" exitCode=0 Oct 07 23:44:00 crc kubenswrapper[4871]: I1007 23:44:00.702304 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" event={"ID":"00e42cab-5504-46f1-a7ad-a74f4007bfef","Type":"ContainerDied","Data":"8885cb7238f5b3a02820c42f2a0544f7b5ff9f92b5de2a2fab0f68d83810b600"} Oct 07 23:44:00 crc kubenswrapper[4871]: I1007 23:44:00.704013 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6","Type":"ContainerStarted","Data":"e020cbc38f62809ceda30ee28b1b9e3c053ec981ebb85de2d43ab9d9a309ffdf"} Oct 07 23:44:00 crc kubenswrapper[4871]: I1007 23:44:00.713165 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:44:01 crc kubenswrapper[4871]: I1007 23:44:01.226027 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" podUID="00e42cab-5504-46f1-a7ad-a74f4007bfef" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.44:5353: connect: connection refused" Oct 07 23:44:01 crc kubenswrapper[4871]: I1007 23:44:01.895782 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.013222 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-sb\") pod \"00e42cab-5504-46f1-a7ad-a74f4007bfef\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.013353 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-dns-svc\") pod \"00e42cab-5504-46f1-a7ad-a74f4007bfef\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.013458 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdvwm\" (UniqueName: \"kubernetes.io/projected/00e42cab-5504-46f1-a7ad-a74f4007bfef-kube-api-access-jdvwm\") pod \"00e42cab-5504-46f1-a7ad-a74f4007bfef\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.013486 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-config\") pod \"00e42cab-5504-46f1-a7ad-a74f4007bfef\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.014411 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-nb\") pod \"00e42cab-5504-46f1-a7ad-a74f4007bfef\" (UID: \"00e42cab-5504-46f1-a7ad-a74f4007bfef\") " Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.018381 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00e42cab-5504-46f1-a7ad-a74f4007bfef-kube-api-access-jdvwm" (OuterVolumeSpecName: "kube-api-access-jdvwm") pod "00e42cab-5504-46f1-a7ad-a74f4007bfef" (UID: "00e42cab-5504-46f1-a7ad-a74f4007bfef"). InnerVolumeSpecName "kube-api-access-jdvwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.034332 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.068621 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "00e42cab-5504-46f1-a7ad-a74f4007bfef" (UID: "00e42cab-5504-46f1-a7ad-a74f4007bfef"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.068730 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "00e42cab-5504-46f1-a7ad-a74f4007bfef" (UID: "00e42cab-5504-46f1-a7ad-a74f4007bfef"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.077459 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-config" (OuterVolumeSpecName: "config") pod "00e42cab-5504-46f1-a7ad-a74f4007bfef" (UID: "00e42cab-5504-46f1-a7ad-a74f4007bfef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.081290 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00e42cab-5504-46f1-a7ad-a74f4007bfef" (UID: "00e42cab-5504-46f1-a7ad-a74f4007bfef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.117923 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.117986 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.118004 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdvwm\" (UniqueName: \"kubernetes.io/projected/00e42cab-5504-46f1-a7ad-a74f4007bfef-kube-api-access-jdvwm\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.118021 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.118035 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e42cab-5504-46f1-a7ad-a74f4007bfef-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.503173 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.626419 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-combined-ca-bundle\") pod \"be095c81-d042-480a-9000-f77699dcfdd6\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.626578 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-config-data\") pod \"be095c81-d042-480a-9000-f77699dcfdd6\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.626644 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpm5m\" (UniqueName: \"kubernetes.io/projected/be095c81-d042-480a-9000-f77699dcfdd6-kube-api-access-zpm5m\") pod \"be095c81-d042-480a-9000-f77699dcfdd6\" (UID: \"be095c81-d042-480a-9000-f77699dcfdd6\") " Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.637051 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be095c81-d042-480a-9000-f77699dcfdd6-kube-api-access-zpm5m" (OuterVolumeSpecName: "kube-api-access-zpm5m") pod "be095c81-d042-480a-9000-f77699dcfdd6" (UID: "be095c81-d042-480a-9000-f77699dcfdd6"). InnerVolumeSpecName "kube-api-access-zpm5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.664075 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-config-data" (OuterVolumeSpecName: "config-data") pod "be095c81-d042-480a-9000-f77699dcfdd6" (UID: "be095c81-d042-480a-9000-f77699dcfdd6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.679695 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be095c81-d042-480a-9000-f77699dcfdd6" (UID: "be095c81-d042-480a-9000-f77699dcfdd6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.712613 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-m5lll"] Oct 07 23:44:02 crc kubenswrapper[4871]: E1007 23:44:02.712993 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be095c81-d042-480a-9000-f77699dcfdd6" containerName="nova-scheduler-scheduler" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.713005 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="be095c81-d042-480a-9000-f77699dcfdd6" containerName="nova-scheduler-scheduler" Oct 07 23:44:02 crc kubenswrapper[4871]: E1007 23:44:02.713015 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e42cab-5504-46f1-a7ad-a74f4007bfef" containerName="init" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.713020 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e42cab-5504-46f1-a7ad-a74f4007bfef" containerName="init" Oct 07 23:44:02 crc kubenswrapper[4871]: E1007 23:44:02.713037 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e42cab-5504-46f1-a7ad-a74f4007bfef" containerName="dnsmasq-dns" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.713043 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e42cab-5504-46f1-a7ad-a74f4007bfef" containerName="dnsmasq-dns" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.713202 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="be095c81-d042-480a-9000-f77699dcfdd6" containerName="nova-scheduler-scheduler" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.713211 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e42cab-5504-46f1-a7ad-a74f4007bfef" containerName="dnsmasq-dns" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.713772 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.719295 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.719608 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.720728 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-m5lll"] Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.731068 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.731157 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpm5m\" (UniqueName: \"kubernetes.io/projected/be095c81-d042-480a-9000-f77699dcfdd6-kube-api-access-zpm5m\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.731180 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be095c81-d042-480a-9000-f77699dcfdd6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.736678 4871 generic.go:334] "Generic (PLEG): container finished" podID="be095c81-d042-480a-9000-f77699dcfdd6" containerID="c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc" exitCode=0 Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.736742 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be095c81-d042-480a-9000-f77699dcfdd6","Type":"ContainerDied","Data":"c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc"} Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.736769 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be095c81-d042-480a-9000-f77699dcfdd6","Type":"ContainerDied","Data":"fc9c93083c3b138ce1da50d314650a44a756f4fca0e8c34a2f128bae612546e3"} Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.736788 4871 scope.go:117] "RemoveContainer" containerID="c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.736984 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.749601 4871 generic.go:334] "Generic (PLEG): container finished" podID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerID="9b45aebedb9fa68e961cc69c307ec232bc2dced405c75f6cda6a0bf7d37e50ea" exitCode=0 Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.749657 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-644z4" event={"ID":"8a6ef7d8-24dd-4edb-b492-b430adf2b67f","Type":"ContainerDied","Data":"9b45aebedb9fa68e961cc69c307ec232bc2dced405c75f6cda6a0bf7d37e50ea"} Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.755162 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76f72129-cdf2-48c2-92e1-555817d6e25b","Type":"ContainerStarted","Data":"5bb6e8ae1eb4290f840107e66a6abc36de3b4a7dcab10f622e5619e64139e9c5"} Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.758386 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" event={"ID":"00e42cab-5504-46f1-a7ad-a74f4007bfef","Type":"ContainerDied","Data":"4e1a7528e05d9edf577bcee741a164d991f1b2bb8ba96568bbd4998296a41ed0"} Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.758502 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b67b947d7-cks8k" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.761025 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6","Type":"ContainerStarted","Data":"49a9e8a177238115f373d908ceb2588e15819b9848026176315cf2c5df2eed61"} Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.782554 4871 scope.go:117] "RemoveContainer" containerID="c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc" Oct 07 23:44:02 crc kubenswrapper[4871]: E1007 23:44:02.783475 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc\": container with ID starting with c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc not found: ID does not exist" containerID="c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.783579 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc"} err="failed to get container status \"c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc\": rpc error: code = NotFound desc = could not find container \"c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc\": container with ID starting with c7f322b653eac3d044214ac166ca067c8288be273a9b4b66e6663298d779c6cc not found: ID does not exist" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.785998 4871 scope.go:117] "RemoveContainer" containerID="8885cb7238f5b3a02820c42f2a0544f7b5ff9f92b5de2a2fab0f68d83810b600" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.785995 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.785982108 podStartE2EDuration="4.785982108s" podCreationTimestamp="2025-10-07 23:43:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:44:02.782983709 +0000 UTC m=+5716.585681792" watchObservedRunningTime="2025-10-07 23:44:02.785982108 +0000 UTC m=+5716.588680181" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.803132 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.80310035 podStartE2EDuration="4.80310035s" podCreationTimestamp="2025-10-07 23:43:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:44:02.802630007 +0000 UTC m=+5716.605328080" watchObservedRunningTime="2025-10-07 23:44:02.80310035 +0000 UTC m=+5716.605798423" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.809682 4871 scope.go:117] "RemoveContainer" containerID="b493fd5137e12dc4e6666e5e6543a533c297a9bf0a5c63089cf0c03c155b9869" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.832869 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-scripts\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.833018 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-config-data\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.833116 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6ng6\" (UniqueName: \"kubernetes.io/projected/bcae407f-430e-49ad-833e-c027eeabb01a-kube-api-access-f6ng6\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.833162 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.833628 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.845627 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.865809 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b67b947d7-cks8k"] Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.891873 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.893514 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.906613 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.918205 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b67b947d7-cks8k"] Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.937922 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6ng6\" (UniqueName: \"kubernetes.io/projected/bcae407f-430e-49ad-833e-c027eeabb01a-kube-api-access-f6ng6\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.937986 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.938030 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-scripts\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.938145 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-config-data\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.942386 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.970465 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6ng6\" (UniqueName: \"kubernetes.io/projected/bcae407f-430e-49ad-833e-c027eeabb01a-kube-api-access-f6ng6\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.971547 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-scripts\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.976487 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-config-data\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:02 crc kubenswrapper[4871]: I1007 23:44:02.980532 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-m5lll\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.018117 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00e42cab-5504-46f1-a7ad-a74f4007bfef" path="/var/lib/kubelet/pods/00e42cab-5504-46f1-a7ad-a74f4007bfef/volumes" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.035008 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be095c81-d042-480a-9000-f77699dcfdd6" path="/var/lib/kubelet/pods/be095c81-d042-480a-9000-f77699dcfdd6/volumes" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.042474 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzmhl\" (UniqueName: \"kubernetes.io/projected/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-kube-api-access-vzmhl\") pod \"nova-scheduler-0\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.042584 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.042674 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-config-data\") pod \"nova-scheduler-0\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.077120 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.146496 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-config-data\") pod \"nova-scheduler-0\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.146631 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzmhl\" (UniqueName: \"kubernetes.io/projected/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-kube-api-access-vzmhl\") pod \"nova-scheduler-0\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.146735 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.152586 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.152620 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-config-data\") pod \"nova-scheduler-0\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.166215 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzmhl\" (UniqueName: \"kubernetes.io/projected/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-kube-api-access-vzmhl\") pod \"nova-scheduler-0\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.252299 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.509923 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-m5lll"] Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.675026 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:03 crc kubenswrapper[4871]: W1007 23:44:03.676749 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7998f1d_8449_4bcb_b7e5_0ef9d66f535e.slice/crio-ab216bd49ca7c433cabd8a4b2f5511956d7a2aec25f86098133eb9192a517495 WatchSource:0}: Error finding container ab216bd49ca7c433cabd8a4b2f5511956d7a2aec25f86098133eb9192a517495: Status 404 returned error can't find the container with id ab216bd49ca7c433cabd8a4b2f5511956d7a2aec25f86098133eb9192a517495 Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.774301 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-644z4" event={"ID":"8a6ef7d8-24dd-4edb-b492-b430adf2b67f","Type":"ContainerStarted","Data":"4eaf07e5dfb893f8de9799688857cbf05de76941207b3d75ec95f1c028dd99f4"} Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.778853 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e","Type":"ContainerStarted","Data":"ab216bd49ca7c433cabd8a4b2f5511956d7a2aec25f86098133eb9192a517495"} Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.780219 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m5lll" event={"ID":"bcae407f-430e-49ad-833e-c027eeabb01a","Type":"ContainerStarted","Data":"3104f2ccfd5f5891af2334afadb5fedde54e4afa72e94fea9cbff6941bc50f09"} Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.780243 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m5lll" event={"ID":"bcae407f-430e-49ad-833e-c027eeabb01a","Type":"ContainerStarted","Data":"b5c13e73c12a046f3b9e7e3e1c282ed818916463824cd13a898cd42500daf87f"} Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.802396 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-644z4" podStartSLOduration=2.078392871 podStartE2EDuration="8.802374709s" podCreationTimestamp="2025-10-07 23:43:55 +0000 UTC" firstStartedPulling="2025-10-07 23:43:56.599689624 +0000 UTC m=+5710.402387737" lastFinishedPulling="2025-10-07 23:44:03.323671502 +0000 UTC m=+5717.126369575" observedRunningTime="2025-10-07 23:44:03.798259531 +0000 UTC m=+5717.600957604" watchObservedRunningTime="2025-10-07 23:44:03.802374709 +0000 UTC m=+5717.605072772" Oct 07 23:44:03 crc kubenswrapper[4871]: I1007 23:44:03.817690 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-m5lll" podStartSLOduration=1.817673263 podStartE2EDuration="1.817673263s" podCreationTimestamp="2025-10-07 23:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:44:03.811372787 +0000 UTC m=+5717.614070860" watchObservedRunningTime="2025-10-07 23:44:03.817673263 +0000 UTC m=+5717.620371336" Oct 07 23:44:04 crc kubenswrapper[4871]: I1007 23:44:04.145460 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 23:44:04 crc kubenswrapper[4871]: I1007 23:44:04.145528 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 23:44:04 crc kubenswrapper[4871]: I1007 23:44:04.799256 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e","Type":"ContainerStarted","Data":"b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49"} Oct 07 23:44:04 crc kubenswrapper[4871]: I1007 23:44:04.835732 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.835479513 podStartE2EDuration="2.835479513s" podCreationTimestamp="2025-10-07 23:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:44:04.829435403 +0000 UTC m=+5718.632133476" watchObservedRunningTime="2025-10-07 23:44:04.835479513 +0000 UTC m=+5718.638177616" Oct 07 23:44:05 crc kubenswrapper[4871]: I1007 23:44:05.619323 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:44:05 crc kubenswrapper[4871]: I1007 23:44:05.619648 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:44:05 crc kubenswrapper[4871]: I1007 23:44:05.707322 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:44:05 crc kubenswrapper[4871]: I1007 23:44:05.983309 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:44:05 crc kubenswrapper[4871]: E1007 23:44:05.983642 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:44:08 crc kubenswrapper[4871]: I1007 23:44:08.253625 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 23:44:08 crc kubenswrapper[4871]: I1007 23:44:08.857675 4871 generic.go:334] "Generic (PLEG): container finished" podID="bcae407f-430e-49ad-833e-c027eeabb01a" containerID="3104f2ccfd5f5891af2334afadb5fedde54e4afa72e94fea9cbff6941bc50f09" exitCode=0 Oct 07 23:44:08 crc kubenswrapper[4871]: I1007 23:44:08.857740 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m5lll" event={"ID":"bcae407f-430e-49ad-833e-c027eeabb01a","Type":"ContainerDied","Data":"3104f2ccfd5f5891af2334afadb5fedde54e4afa72e94fea9cbff6941bc50f09"} Oct 07 23:44:09 crc kubenswrapper[4871]: I1007 23:44:09.100064 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 23:44:09 crc kubenswrapper[4871]: I1007 23:44:09.100137 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 23:44:09 crc kubenswrapper[4871]: I1007 23:44:09.146247 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 23:44:09 crc kubenswrapper[4871]: I1007 23:44:09.146348 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.182148 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.64:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.182161 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.64:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.224035 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.65:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.257523 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.265153 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.65:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.396229 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-scripts\") pod \"bcae407f-430e-49ad-833e-c027eeabb01a\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.396362 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-config-data\") pod \"bcae407f-430e-49ad-833e-c027eeabb01a\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.396409 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-combined-ca-bundle\") pod \"bcae407f-430e-49ad-833e-c027eeabb01a\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.396485 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6ng6\" (UniqueName: \"kubernetes.io/projected/bcae407f-430e-49ad-833e-c027eeabb01a-kube-api-access-f6ng6\") pod \"bcae407f-430e-49ad-833e-c027eeabb01a\" (UID: \"bcae407f-430e-49ad-833e-c027eeabb01a\") " Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.403901 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-scripts" (OuterVolumeSpecName: "scripts") pod "bcae407f-430e-49ad-833e-c027eeabb01a" (UID: "bcae407f-430e-49ad-833e-c027eeabb01a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.403971 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcae407f-430e-49ad-833e-c027eeabb01a-kube-api-access-f6ng6" (OuterVolumeSpecName: "kube-api-access-f6ng6") pod "bcae407f-430e-49ad-833e-c027eeabb01a" (UID: "bcae407f-430e-49ad-833e-c027eeabb01a"). InnerVolumeSpecName "kube-api-access-f6ng6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.441973 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-config-data" (OuterVolumeSpecName: "config-data") pod "bcae407f-430e-49ad-833e-c027eeabb01a" (UID: "bcae407f-430e-49ad-833e-c027eeabb01a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.448938 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcae407f-430e-49ad-833e-c027eeabb01a" (UID: "bcae407f-430e-49ad-833e-c027eeabb01a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.498728 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.498761 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.498770 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcae407f-430e-49ad-833e-c027eeabb01a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.498781 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6ng6\" (UniqueName: \"kubernetes.io/projected/bcae407f-430e-49ad-833e-c027eeabb01a-kube-api-access-f6ng6\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.883167 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-m5lll" event={"ID":"bcae407f-430e-49ad-833e-c027eeabb01a","Type":"ContainerDied","Data":"b5c13e73c12a046f3b9e7e3e1c282ed818916463824cd13a898cd42500daf87f"} Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.883536 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5c13e73c12a046f3b9e7e3e1c282ed818916463824cd13a898cd42500daf87f" Oct 07 23:44:10 crc kubenswrapper[4871]: I1007 23:44:10.883220 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-m5lll" Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.077920 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.078147 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d7998f1d-8449-4bcb-b7e5-0ef9d66f535e" containerName="nova-scheduler-scheduler" containerID="cri-o://b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49" gracePeriod=30 Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.090259 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.090454 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-log" containerID="cri-o://e020cbc38f62809ceda30ee28b1b9e3c053ec981ebb85de2d43ab9d9a309ffdf" gracePeriod=30 Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.091015 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-api" containerID="cri-o://49a9e8a177238115f373d908ceb2588e15819b9848026176315cf2c5df2eed61" gracePeriod=30 Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.117657 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.121746 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-log" containerID="cri-o://042233c1ac1f908b0a89151b3a0863aa28ba0813996b31e2fa2e7dc378a11212" gracePeriod=30 Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.121823 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-metadata" containerID="cri-o://5bb6e8ae1eb4290f840107e66a6abc36de3b4a7dcab10f622e5619e64139e9c5" gracePeriod=30 Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.894613 4871 generic.go:334] "Generic (PLEG): container finished" podID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerID="042233c1ac1f908b0a89151b3a0863aa28ba0813996b31e2fa2e7dc378a11212" exitCode=143 Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.894733 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76f72129-cdf2-48c2-92e1-555817d6e25b","Type":"ContainerDied","Data":"042233c1ac1f908b0a89151b3a0863aa28ba0813996b31e2fa2e7dc378a11212"} Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.897097 4871 generic.go:334] "Generic (PLEG): container finished" podID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerID="e020cbc38f62809ceda30ee28b1b9e3c053ec981ebb85de2d43ab9d9a309ffdf" exitCode=143 Oct 07 23:44:11 crc kubenswrapper[4871]: I1007 23:44:11.897136 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6","Type":"ContainerDied","Data":"e020cbc38f62809ceda30ee28b1b9e3c053ec981ebb85de2d43ab9d9a309ffdf"} Oct 07 23:44:14 crc kubenswrapper[4871]: I1007 23:44:14.942528 4871 generic.go:334] "Generic (PLEG): container finished" podID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerID="5bb6e8ae1eb4290f840107e66a6abc36de3b4a7dcab10f622e5619e64139e9c5" exitCode=0 Oct 07 23:44:14 crc kubenswrapper[4871]: I1007 23:44:14.942676 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76f72129-cdf2-48c2-92e1-555817d6e25b","Type":"ContainerDied","Data":"5bb6e8ae1eb4290f840107e66a6abc36de3b4a7dcab10f622e5619e64139e9c5"} Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.042316 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.204219 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f72129-cdf2-48c2-92e1-555817d6e25b-logs\") pod \"76f72129-cdf2-48c2-92e1-555817d6e25b\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.204383 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-config-data\") pod \"76f72129-cdf2-48c2-92e1-555817d6e25b\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.204453 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f4kh\" (UniqueName: \"kubernetes.io/projected/76f72129-cdf2-48c2-92e1-555817d6e25b-kube-api-access-5f4kh\") pod \"76f72129-cdf2-48c2-92e1-555817d6e25b\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.204515 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-combined-ca-bundle\") pod \"76f72129-cdf2-48c2-92e1-555817d6e25b\" (UID: \"76f72129-cdf2-48c2-92e1-555817d6e25b\") " Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.205397 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76f72129-cdf2-48c2-92e1-555817d6e25b-logs" (OuterVolumeSpecName: "logs") pod "76f72129-cdf2-48c2-92e1-555817d6e25b" (UID: "76f72129-cdf2-48c2-92e1-555817d6e25b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.210053 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76f72129-cdf2-48c2-92e1-555817d6e25b-kube-api-access-5f4kh" (OuterVolumeSpecName: "kube-api-access-5f4kh") pod "76f72129-cdf2-48c2-92e1-555817d6e25b" (UID: "76f72129-cdf2-48c2-92e1-555817d6e25b"). InnerVolumeSpecName "kube-api-access-5f4kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.240501 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76f72129-cdf2-48c2-92e1-555817d6e25b" (UID: "76f72129-cdf2-48c2-92e1-555817d6e25b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.247770 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-config-data" (OuterVolumeSpecName: "config-data") pod "76f72129-cdf2-48c2-92e1-555817d6e25b" (UID: "76f72129-cdf2-48c2-92e1-555817d6e25b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.307307 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f4kh\" (UniqueName: \"kubernetes.io/projected/76f72129-cdf2-48c2-92e1-555817d6e25b-kube-api-access-5f4kh\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.307342 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.307354 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f72129-cdf2-48c2-92e1-555817d6e25b-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.307366 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f72129-cdf2-48c2-92e1-555817d6e25b-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.309038 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.409016 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-config-data\") pod \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.409288 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-combined-ca-bundle\") pod \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.409378 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzmhl\" (UniqueName: \"kubernetes.io/projected/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-kube-api-access-vzmhl\") pod \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\" (UID: \"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e\") " Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.415168 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-kube-api-access-vzmhl" (OuterVolumeSpecName: "kube-api-access-vzmhl") pod "d7998f1d-8449-4bcb-b7e5-0ef9d66f535e" (UID: "d7998f1d-8449-4bcb-b7e5-0ef9d66f535e"). InnerVolumeSpecName "kube-api-access-vzmhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.438544 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7998f1d-8449-4bcb-b7e5-0ef9d66f535e" (UID: "d7998f1d-8449-4bcb-b7e5-0ef9d66f535e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.450286 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-config-data" (OuterVolumeSpecName: "config-data") pod "d7998f1d-8449-4bcb-b7e5-0ef9d66f535e" (UID: "d7998f1d-8449-4bcb-b7e5-0ef9d66f535e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.510617 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.510846 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzmhl\" (UniqueName: \"kubernetes.io/projected/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-kube-api-access-vzmhl\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.510946 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.668514 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-644z4" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.956507 4871 generic.go:334] "Generic (PLEG): container finished" podID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerID="49a9e8a177238115f373d908ceb2588e15819b9848026176315cf2c5df2eed61" exitCode=0 Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.956565 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6","Type":"ContainerDied","Data":"49a9e8a177238115f373d908ceb2588e15819b9848026176315cf2c5df2eed61"} Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.956649 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6","Type":"ContainerDied","Data":"fa30f725c0b73dac68fd53bdc17389b383d71b7cd4312a25a7dab36ad0fb5222"} Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.956670 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa30f725c0b73dac68fd53bdc17389b383d71b7cd4312a25a7dab36ad0fb5222" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.958274 4871 generic.go:334] "Generic (PLEG): container finished" podID="d7998f1d-8449-4bcb-b7e5-0ef9d66f535e" containerID="b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49" exitCode=0 Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.958331 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.958350 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e","Type":"ContainerDied","Data":"b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49"} Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.958422 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7998f1d-8449-4bcb-b7e5-0ef9d66f535e","Type":"ContainerDied","Data":"ab216bd49ca7c433cabd8a4b2f5511956d7a2aec25f86098133eb9192a517495"} Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.958455 4871 scope.go:117] "RemoveContainer" containerID="b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49" Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.963754 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76f72129-cdf2-48c2-92e1-555817d6e25b","Type":"ContainerDied","Data":"28ca81145da334b1f25dce453221b2f29524112eb0eddb56ca708106f254a884"} Oct 07 23:44:15 crc kubenswrapper[4871]: I1007 23:44:15.963872 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.012474 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.031526 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.054325 4871 scope.go:117] "RemoveContainer" containerID="b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.054540 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:16 crc kubenswrapper[4871]: E1007 23:44:16.058738 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49\": container with ID starting with b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49 not found: ID does not exist" containerID="b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.058860 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49"} err="failed to get container status \"b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49\": rpc error: code = NotFound desc = could not find container \"b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49\": container with ID starting with b0eca78211f92ace852c6668d92c20baf3271ae101c7dcc85acef11f2e993a49 not found: ID does not exist" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.058924 4871 scope.go:117] "RemoveContainer" containerID="5bb6e8ae1eb4290f840107e66a6abc36de3b4a7dcab10f622e5619e64139e9c5" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.064986 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:16 crc kubenswrapper[4871]: E1007 23:44:16.065458 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7998f1d-8449-4bcb-b7e5-0ef9d66f535e" containerName="nova-scheduler-scheduler" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065479 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7998f1d-8449-4bcb-b7e5-0ef9d66f535e" containerName="nova-scheduler-scheduler" Oct 07 23:44:16 crc kubenswrapper[4871]: E1007 23:44:16.065491 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-log" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065499 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-log" Oct 07 23:44:16 crc kubenswrapper[4871]: E1007 23:44:16.065510 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-api" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065516 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-api" Oct 07 23:44:16 crc kubenswrapper[4871]: E1007 23:44:16.065530 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-metadata" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065536 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-metadata" Oct 07 23:44:16 crc kubenswrapper[4871]: E1007 23:44:16.065563 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-log" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065570 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-log" Oct 07 23:44:16 crc kubenswrapper[4871]: E1007 23:44:16.065585 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcae407f-430e-49ad-833e-c027eeabb01a" containerName="nova-manage" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065592 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcae407f-430e-49ad-833e-c027eeabb01a" containerName="nova-manage" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065746 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-metadata" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065770 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7998f1d-8449-4bcb-b7e5-0ef9d66f535e" containerName="nova-scheduler-scheduler" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065778 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-api" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065805 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" containerName="nova-api-log" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065816 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" containerName="nova-metadata-log" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.065829 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcae407f-430e-49ad-833e-c027eeabb01a" containerName="nova-manage" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.066444 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.072603 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.074249 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.103120 4871 scope.go:117] "RemoveContainer" containerID="042233c1ac1f908b0a89151b3a0863aa28ba0813996b31e2fa2e7dc378a11212" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.117557 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.128210 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-logs\") pod \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.128269 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-config-data\") pod \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.128410 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhq55\" (UniqueName: \"kubernetes.io/projected/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-kube-api-access-jhq55\") pod \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.128442 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-combined-ca-bundle\") pod \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\" (UID: \"58a418f3-58f8-4e04-b3d7-1b7f7543b1a6\") " Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.129278 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.130173 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-logs" (OuterVolumeSpecName: "logs") pod "58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" (UID: "58a418f3-58f8-4e04-b3d7-1b7f7543b1a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.131930 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-kube-api-access-jhq55" (OuterVolumeSpecName: "kube-api-access-jhq55") pod "58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" (UID: "58a418f3-58f8-4e04-b3d7-1b7f7543b1a6"). InnerVolumeSpecName "kube-api-access-jhq55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.162629 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.165637 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.165754 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.168777 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.172052 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-config-data" (OuterVolumeSpecName: "config-data") pod "58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" (UID: "58a418f3-58f8-4e04-b3d7-1b7f7543b1a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.185234 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" (UID: "58a418f3-58f8-4e04-b3d7-1b7f7543b1a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.230943 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.231285 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4brhs\" (UniqueName: \"kubernetes.io/projected/5e592305-57bf-441b-926e-0217c88ab964-kube-api-access-4brhs\") pod \"nova-scheduler-0\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.231415 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-config-data\") pod \"nova-scheduler-0\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.231620 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.231648 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhq55\" (UniqueName: \"kubernetes.io/projected/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-kube-api-access-jhq55\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.231663 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.231676 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.332818 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4brhs\" (UniqueName: \"kubernetes.io/projected/5e592305-57bf-441b-926e-0217c88ab964-kube-api-access-4brhs\") pod \"nova-scheduler-0\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.334007 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-config-data\") pod \"nova-scheduler-0\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.334204 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvkqh\" (UniqueName: \"kubernetes.io/projected/f23c5ed1-6597-4d11-9270-6c0428a37d42-kube-api-access-gvkqh\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.334479 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.334532 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.334597 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-config-data\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.334644 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f23c5ed1-6597-4d11-9270-6c0428a37d42-logs\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.338029 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-config-data\") pod \"nova-scheduler-0\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.338187 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.349253 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4brhs\" (UniqueName: \"kubernetes.io/projected/5e592305-57bf-441b-926e-0217c88ab964-kube-api-access-4brhs\") pod \"nova-scheduler-0\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.402647 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.435633 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvkqh\" (UniqueName: \"kubernetes.io/projected/f23c5ed1-6597-4d11-9270-6c0428a37d42-kube-api-access-gvkqh\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.435866 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.435926 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-config-data\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.435981 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f23c5ed1-6597-4d11-9270-6c0428a37d42-logs\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.436357 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f23c5ed1-6597-4d11-9270-6c0428a37d42-logs\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.440845 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.441538 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-config-data\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.451937 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvkqh\" (UniqueName: \"kubernetes.io/projected/f23c5ed1-6597-4d11-9270-6c0428a37d42-kube-api-access-gvkqh\") pod \"nova-metadata-0\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.513122 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.888484 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.977676 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5e592305-57bf-441b-926e-0217c88ab964","Type":"ContainerStarted","Data":"b0cd581e551a5710791dbe47f87497cbb922b52f595ebd96cabe01be0c8f37fb"} Oct 07 23:44:16 crc kubenswrapper[4871]: I1007 23:44:16.979575 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.002932 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76f72129-cdf2-48c2-92e1-555817d6e25b" path="/var/lib/kubelet/pods/76f72129-cdf2-48c2-92e1-555817d6e25b/volumes" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.003991 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7998f1d-8449-4bcb-b7e5-0ef9d66f535e" path="/var/lib/kubelet/pods/d7998f1d-8449-4bcb-b7e5-0ef9d66f535e/volumes" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.010496 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.067517 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.076656 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.084731 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.087182 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.092423 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.093347 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.251909 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-config-data\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.251972 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf65k\" (UniqueName: \"kubernetes.io/projected/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-kube-api-access-zf65k\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.252775 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-logs\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.252961 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.354354 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-config-data\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.354709 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf65k\" (UniqueName: \"kubernetes.io/projected/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-kube-api-access-zf65k\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.355022 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-logs\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.355623 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-logs\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.359424 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.362333 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-config-data\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.363511 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.379001 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf65k\" (UniqueName: \"kubernetes.io/projected/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-kube-api-access-zf65k\") pod \"nova-api-0\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.459676 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.983188 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:44:17 crc kubenswrapper[4871]: E1007 23:44:17.984147 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.990511 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.993561 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5e592305-57bf-441b-926e-0217c88ab964","Type":"ContainerStarted","Data":"ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b"} Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.995183 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f23c5ed1-6597-4d11-9270-6c0428a37d42","Type":"ContainerStarted","Data":"f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67"} Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.995232 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f23c5ed1-6597-4d11-9270-6c0428a37d42","Type":"ContainerStarted","Data":"38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558"} Oct 07 23:44:17 crc kubenswrapper[4871]: I1007 23:44:17.995255 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f23c5ed1-6597-4d11-9270-6c0428a37d42","Type":"ContainerStarted","Data":"5ed53e44433f6372af212c04d09fbbc0ae3d8e2bbd9d7a301350f160836609c0"} Oct 07 23:44:17 crc kubenswrapper[4871]: W1007 23:44:17.997406 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e5ea6bf_7e9c_4dd2_9c9c_6c98b9e7280c.slice/crio-3e7f63389a5f6c0c06f8ed5ea9a4f5d30e654898d9eebf7428f4c5851d4ff5da WatchSource:0}: Error finding container 3e7f63389a5f6c0c06f8ed5ea9a4f5d30e654898d9eebf7428f4c5851d4ff5da: Status 404 returned error can't find the container with id 3e7f63389a5f6c0c06f8ed5ea9a4f5d30e654898d9eebf7428f4c5851d4ff5da Oct 07 23:44:18 crc kubenswrapper[4871]: I1007 23:44:18.017083 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.017066336 podStartE2EDuration="2.017066336s" podCreationTimestamp="2025-10-07 23:44:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:44:18.011501799 +0000 UTC m=+5731.814199912" watchObservedRunningTime="2025-10-07 23:44:18.017066336 +0000 UTC m=+5731.819764409" Oct 07 23:44:18 crc kubenswrapper[4871]: I1007 23:44:18.059039 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.059014143 podStartE2EDuration="2.059014143s" podCreationTimestamp="2025-10-07 23:44:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:44:18.045849015 +0000 UTC m=+5731.848547088" watchObservedRunningTime="2025-10-07 23:44:18.059014143 +0000 UTC m=+5731.861712236" Oct 07 23:44:19 crc kubenswrapper[4871]: I1007 23:44:19.004718 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58a418f3-58f8-4e04-b3d7-1b7f7543b1a6" path="/var/lib/kubelet/pods/58a418f3-58f8-4e04-b3d7-1b7f7543b1a6/volumes" Oct 07 23:44:19 crc kubenswrapper[4871]: I1007 23:44:19.018764 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c","Type":"ContainerStarted","Data":"c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1"} Oct 07 23:44:19 crc kubenswrapper[4871]: I1007 23:44:19.018848 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c","Type":"ContainerStarted","Data":"8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1"} Oct 07 23:44:19 crc kubenswrapper[4871]: I1007 23:44:19.018874 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c","Type":"ContainerStarted","Data":"3e7f63389a5f6c0c06f8ed5ea9a4f5d30e654898d9eebf7428f4c5851d4ff5da"} Oct 07 23:44:19 crc kubenswrapper[4871]: I1007 23:44:19.059387 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.059368722 podStartE2EDuration="2.059368722s" podCreationTimestamp="2025-10-07 23:44:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:44:19.045474655 +0000 UTC m=+5732.848172728" watchObservedRunningTime="2025-10-07 23:44:19.059368722 +0000 UTC m=+5732.862066795" Oct 07 23:44:19 crc kubenswrapper[4871]: I1007 23:44:19.501695 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-644z4"] Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.099141 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5gn7g"] Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.099557 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5gn7g" podUID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerName="registry-server" containerID="cri-o://4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e" gracePeriod=2 Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.524755 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.630355 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-utilities\") pod \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.630494 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lplmt\" (UniqueName: \"kubernetes.io/projected/ac1066e8-f230-4d5f-9a0c-639f2874fab5-kube-api-access-lplmt\") pod \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.630587 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-catalog-content\") pod \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\" (UID: \"ac1066e8-f230-4d5f-9a0c-639f2874fab5\") " Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.631784 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-utilities" (OuterVolumeSpecName: "utilities") pod "ac1066e8-f230-4d5f-9a0c-639f2874fab5" (UID: "ac1066e8-f230-4d5f-9a0c-639f2874fab5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.641710 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac1066e8-f230-4d5f-9a0c-639f2874fab5-kube-api-access-lplmt" (OuterVolumeSpecName: "kube-api-access-lplmt") pod "ac1066e8-f230-4d5f-9a0c-639f2874fab5" (UID: "ac1066e8-f230-4d5f-9a0c-639f2874fab5"). InnerVolumeSpecName "kube-api-access-lplmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.675872 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac1066e8-f230-4d5f-9a0c-639f2874fab5" (UID: "ac1066e8-f230-4d5f-9a0c-639f2874fab5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.732276 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.732308 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1066e8-f230-4d5f-9a0c-639f2874fab5-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:20 crc kubenswrapper[4871]: I1007 23:44:20.732318 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lplmt\" (UniqueName: \"kubernetes.io/projected/ac1066e8-f230-4d5f-9a0c-639f2874fab5-kube-api-access-lplmt\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.043473 4871 generic.go:334] "Generic (PLEG): container finished" podID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerID="4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e" exitCode=0 Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.043530 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gn7g" event={"ID":"ac1066e8-f230-4d5f-9a0c-639f2874fab5","Type":"ContainerDied","Data":"4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e"} Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.043578 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gn7g" event={"ID":"ac1066e8-f230-4d5f-9a0c-639f2874fab5","Type":"ContainerDied","Data":"30d69b02629c93d1b052a960dc03ddea92c245f58f220fc74145ac80aa7a876b"} Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.043606 4871 scope.go:117] "RemoveContainer" containerID="4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.043613 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gn7g" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.071447 4871 scope.go:117] "RemoveContainer" containerID="43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.071684 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5gn7g"] Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.083515 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5gn7g"] Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.092426 4871 scope.go:117] "RemoveContainer" containerID="cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.145772 4871 scope.go:117] "RemoveContainer" containerID="4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e" Oct 07 23:44:21 crc kubenswrapper[4871]: E1007 23:44:21.146395 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e\": container with ID starting with 4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e not found: ID does not exist" containerID="4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.146424 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e"} err="failed to get container status \"4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e\": rpc error: code = NotFound desc = could not find container \"4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e\": container with ID starting with 4af9af4178ecf9c6bbf5fecdca29b51a5c86fc3eab6b75f5b19885f54481257e not found: ID does not exist" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.146446 4871 scope.go:117] "RemoveContainer" containerID="43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28" Oct 07 23:44:21 crc kubenswrapper[4871]: E1007 23:44:21.146716 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28\": container with ID starting with 43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28 not found: ID does not exist" containerID="43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.146738 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28"} err="failed to get container status \"43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28\": rpc error: code = NotFound desc = could not find container \"43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28\": container with ID starting with 43c18d2208ff297415e3bc0d0ffd95d8f54fa5d3669af7bdcf94390e6c5c7f28 not found: ID does not exist" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.146758 4871 scope.go:117] "RemoveContainer" containerID="cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03" Oct 07 23:44:21 crc kubenswrapper[4871]: E1007 23:44:21.147172 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03\": container with ID starting with cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03 not found: ID does not exist" containerID="cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.147212 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03"} err="failed to get container status \"cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03\": rpc error: code = NotFound desc = could not find container \"cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03\": container with ID starting with cf57f5367e86001df63b77a9f828e865029959c0f0df6de048a26f5c610b7e03 not found: ID does not exist" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.403922 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.513765 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 23:44:21 crc kubenswrapper[4871]: I1007 23:44:21.513939 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 23:44:22 crc kubenswrapper[4871]: I1007 23:44:22.998130 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" path="/var/lib/kubelet/pods/ac1066e8-f230-4d5f-9a0c-639f2874fab5/volumes" Oct 07 23:44:26 crc kubenswrapper[4871]: I1007 23:44:26.404121 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 23:44:26 crc kubenswrapper[4871]: I1007 23:44:26.450460 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 23:44:26 crc kubenswrapper[4871]: I1007 23:44:26.513819 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 23:44:26 crc kubenswrapper[4871]: I1007 23:44:26.513879 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 23:44:27 crc kubenswrapper[4871]: I1007 23:44:27.148665 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 23:44:27 crc kubenswrapper[4871]: I1007 23:44:27.460303 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 23:44:27 crc kubenswrapper[4871]: I1007 23:44:27.460656 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 23:44:27 crc kubenswrapper[4871]: I1007 23:44:27.596160 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:44:27 crc kubenswrapper[4871]: I1007 23:44:27.596201 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:44:28 crc kubenswrapper[4871]: I1007 23:44:28.543012 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:44:28 crc kubenswrapper[4871]: I1007 23:44:28.543505 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:44:29 crc kubenswrapper[4871]: I1007 23:44:29.984359 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:44:29 crc kubenswrapper[4871]: E1007 23:44:29.984886 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:44:36 crc kubenswrapper[4871]: I1007 23:44:36.515456 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 23:44:36 crc kubenswrapper[4871]: I1007 23:44:36.517775 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 23:44:36 crc kubenswrapper[4871]: I1007 23:44:36.518704 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 23:44:37 crc kubenswrapper[4871]: I1007 23:44:37.253199 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 23:44:37 crc kubenswrapper[4871]: I1007 23:44:37.466418 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 23:44:37 crc kubenswrapper[4871]: I1007 23:44:37.466759 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 23:44:37 crc kubenswrapper[4871]: I1007 23:44:37.467329 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 23:44:37 crc kubenswrapper[4871]: I1007 23:44:37.471394 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.262912 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.267639 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.531067 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b4b8d7867-dqbll"] Oct 07 23:44:38 crc kubenswrapper[4871]: E1007 23:44:38.531715 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerName="extract-utilities" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.531732 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerName="extract-utilities" Oct 07 23:44:38 crc kubenswrapper[4871]: E1007 23:44:38.531748 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerName="registry-server" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.531755 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerName="registry-server" Oct 07 23:44:38 crc kubenswrapper[4871]: E1007 23:44:38.531777 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerName="extract-content" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.531785 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerName="extract-content" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.532011 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac1066e8-f230-4d5f-9a0c-639f2874fab5" containerName="registry-server" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.532980 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.542876 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b4b8d7867-dqbll"] Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.710927 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-dns-svc\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.711133 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-nb\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.711309 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-config\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.711399 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-sb\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.711559 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq7gm\" (UniqueName: \"kubernetes.io/projected/20c7f7e8-8bea-4384-a21c-4ed818883882-kube-api-access-mq7gm\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.813250 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq7gm\" (UniqueName: \"kubernetes.io/projected/20c7f7e8-8bea-4384-a21c-4ed818883882-kube-api-access-mq7gm\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.813356 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-dns-svc\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.813425 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-nb\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.813477 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-config\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.813510 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-sb\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.814401 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-sb\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.814430 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-config\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.814673 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-nb\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.815009 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-dns-svc\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.837419 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq7gm\" (UniqueName: \"kubernetes.io/projected/20c7f7e8-8bea-4384-a21c-4ed818883882-kube-api-access-mq7gm\") pod \"dnsmasq-dns-b4b8d7867-dqbll\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:38 crc kubenswrapper[4871]: I1007 23:44:38.895428 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:39 crc kubenswrapper[4871]: I1007 23:44:39.367845 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b4b8d7867-dqbll"] Oct 07 23:44:40 crc kubenswrapper[4871]: I1007 23:44:40.287553 4871 generic.go:334] "Generic (PLEG): container finished" podID="20c7f7e8-8bea-4384-a21c-4ed818883882" containerID="b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619" exitCode=0 Oct 07 23:44:40 crc kubenswrapper[4871]: I1007 23:44:40.287684 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" event={"ID":"20c7f7e8-8bea-4384-a21c-4ed818883882","Type":"ContainerDied","Data":"b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619"} Oct 07 23:44:40 crc kubenswrapper[4871]: I1007 23:44:40.288145 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" event={"ID":"20c7f7e8-8bea-4384-a21c-4ed818883882","Type":"ContainerStarted","Data":"c58ec01186372f44c2e24edf2b6591ad454e2a66715b4fde0358a3fbd4df6777"} Oct 07 23:44:41 crc kubenswrapper[4871]: I1007 23:44:41.301082 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" event={"ID":"20c7f7e8-8bea-4384-a21c-4ed818883882","Type":"ContainerStarted","Data":"57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a"} Oct 07 23:44:41 crc kubenswrapper[4871]: I1007 23:44:41.301651 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:41 crc kubenswrapper[4871]: I1007 23:44:41.328880 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" podStartSLOduration=3.3288563890000002 podStartE2EDuration="3.328856389s" podCreationTimestamp="2025-10-07 23:44:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:44:41.327390651 +0000 UTC m=+5755.130088734" watchObservedRunningTime="2025-10-07 23:44:41.328856389 +0000 UTC m=+5755.131554492" Oct 07 23:44:44 crc kubenswrapper[4871]: I1007 23:44:44.983330 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:44:44 crc kubenswrapper[4871]: E1007 23:44:44.984036 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:44:48 crc kubenswrapper[4871]: I1007 23:44:48.898051 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.019471 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d6b768dc-jcjnc"] Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.023245 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" podUID="f2976020-e44a-44d7-890a-4dc6806d222a" containerName="dnsmasq-dns" containerID="cri-o://e5bb5fed9ae22642b24d7397e58b83db9bd31e81f93799c2e1551bb3ff7d34ef" gracePeriod=10 Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.384678 4871 generic.go:334] "Generic (PLEG): container finished" podID="f2976020-e44a-44d7-890a-4dc6806d222a" containerID="e5bb5fed9ae22642b24d7397e58b83db9bd31e81f93799c2e1551bb3ff7d34ef" exitCode=0 Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.384734 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" event={"ID":"f2976020-e44a-44d7-890a-4dc6806d222a","Type":"ContainerDied","Data":"e5bb5fed9ae22642b24d7397e58b83db9bd31e81f93799c2e1551bb3ff7d34ef"} Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.502044 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.675954 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-sb\") pod \"f2976020-e44a-44d7-890a-4dc6806d222a\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.676017 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-nb\") pod \"f2976020-e44a-44d7-890a-4dc6806d222a\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.676123 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-config\") pod \"f2976020-e44a-44d7-890a-4dc6806d222a\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.676264 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpjkv\" (UniqueName: \"kubernetes.io/projected/f2976020-e44a-44d7-890a-4dc6806d222a-kube-api-access-vpjkv\") pod \"f2976020-e44a-44d7-890a-4dc6806d222a\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.676332 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-dns-svc\") pod \"f2976020-e44a-44d7-890a-4dc6806d222a\" (UID: \"f2976020-e44a-44d7-890a-4dc6806d222a\") " Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.694209 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2976020-e44a-44d7-890a-4dc6806d222a-kube-api-access-vpjkv" (OuterVolumeSpecName: "kube-api-access-vpjkv") pod "f2976020-e44a-44d7-890a-4dc6806d222a" (UID: "f2976020-e44a-44d7-890a-4dc6806d222a"). InnerVolumeSpecName "kube-api-access-vpjkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.717617 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f2976020-e44a-44d7-890a-4dc6806d222a" (UID: "f2976020-e44a-44d7-890a-4dc6806d222a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.720154 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-config" (OuterVolumeSpecName: "config") pod "f2976020-e44a-44d7-890a-4dc6806d222a" (UID: "f2976020-e44a-44d7-890a-4dc6806d222a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.747460 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f2976020-e44a-44d7-890a-4dc6806d222a" (UID: "f2976020-e44a-44d7-890a-4dc6806d222a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.753622 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f2976020-e44a-44d7-890a-4dc6806d222a" (UID: "f2976020-e44a-44d7-890a-4dc6806d222a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.777973 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.778690 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.778839 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.778878 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2976020-e44a-44d7-890a-4dc6806d222a-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:49 crc kubenswrapper[4871]: I1007 23:44:49.778911 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpjkv\" (UniqueName: \"kubernetes.io/projected/f2976020-e44a-44d7-890a-4dc6806d222a-kube-api-access-vpjkv\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:50 crc kubenswrapper[4871]: I1007 23:44:50.400244 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" event={"ID":"f2976020-e44a-44d7-890a-4dc6806d222a","Type":"ContainerDied","Data":"fc2600253e49b6ccec2659681468c647502eede9639027041f005cb041e66e17"} Oct 07 23:44:50 crc kubenswrapper[4871]: I1007 23:44:50.400308 4871 scope.go:117] "RemoveContainer" containerID="e5bb5fed9ae22642b24d7397e58b83db9bd31e81f93799c2e1551bb3ff7d34ef" Oct 07 23:44:50 crc kubenswrapper[4871]: I1007 23:44:50.400315 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6b768dc-jcjnc" Oct 07 23:44:50 crc kubenswrapper[4871]: I1007 23:44:50.447669 4871 scope.go:117] "RemoveContainer" containerID="7e654d77532b9ec11568b64fb8c2d71d4685ac1b9b4d4a1d8794ec17fc62f32b" Oct 07 23:44:50 crc kubenswrapper[4871]: I1007 23:44:50.452095 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d6b768dc-jcjnc"] Oct 07 23:44:50 crc kubenswrapper[4871]: I1007 23:44:50.478223 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d6b768dc-jcjnc"] Oct 07 23:44:50 crc kubenswrapper[4871]: I1007 23:44:50.994442 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2976020-e44a-44d7-890a-4dc6806d222a" path="/var/lib/kubelet/pods/f2976020-e44a-44d7-890a-4dc6806d222a/volumes" Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.477210 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-gtk4b"] Oct 07 23:44:52 crc kubenswrapper[4871]: E1007 23:44:52.477674 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2976020-e44a-44d7-890a-4dc6806d222a" containerName="init" Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.477692 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2976020-e44a-44d7-890a-4dc6806d222a" containerName="init" Oct 07 23:44:52 crc kubenswrapper[4871]: E1007 23:44:52.477708 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2976020-e44a-44d7-890a-4dc6806d222a" containerName="dnsmasq-dns" Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.477717 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2976020-e44a-44d7-890a-4dc6806d222a" containerName="dnsmasq-dns" Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.477991 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2976020-e44a-44d7-890a-4dc6806d222a" containerName="dnsmasq-dns" Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.478762 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-gtk4b" Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.493205 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-gtk4b"] Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.636558 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9rwp\" (UniqueName: \"kubernetes.io/projected/d021d8ab-b7ae-4136-a946-2729bdd971b9-kube-api-access-k9rwp\") pod \"cinder-db-create-gtk4b\" (UID: \"d021d8ab-b7ae-4136-a946-2729bdd971b9\") " pod="openstack/cinder-db-create-gtk4b" Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.738924 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9rwp\" (UniqueName: \"kubernetes.io/projected/d021d8ab-b7ae-4136-a946-2729bdd971b9-kube-api-access-k9rwp\") pod \"cinder-db-create-gtk4b\" (UID: \"d021d8ab-b7ae-4136-a946-2729bdd971b9\") " pod="openstack/cinder-db-create-gtk4b" Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.760526 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9rwp\" (UniqueName: \"kubernetes.io/projected/d021d8ab-b7ae-4136-a946-2729bdd971b9-kube-api-access-k9rwp\") pod \"cinder-db-create-gtk4b\" (UID: \"d021d8ab-b7ae-4136-a946-2729bdd971b9\") " pod="openstack/cinder-db-create-gtk4b" Oct 07 23:44:52 crc kubenswrapper[4871]: I1007 23:44:52.796604 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-gtk4b" Oct 07 23:44:53 crc kubenswrapper[4871]: I1007 23:44:53.294701 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-gtk4b"] Oct 07 23:44:53 crc kubenswrapper[4871]: I1007 23:44:53.433710 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-gtk4b" event={"ID":"d021d8ab-b7ae-4136-a946-2729bdd971b9","Type":"ContainerStarted","Data":"ac6423c8d68a70e2bae2ac41620abb5f2583f84cd4597c81d7e7020002556ad2"} Oct 07 23:44:54 crc kubenswrapper[4871]: I1007 23:44:54.443198 4871 generic.go:334] "Generic (PLEG): container finished" podID="d021d8ab-b7ae-4136-a946-2729bdd971b9" containerID="4a2c2efd8c6101f6ff7df800e3a4a0a8b13b16ff72ae1ec43233a855fb874ea2" exitCode=0 Oct 07 23:44:54 crc kubenswrapper[4871]: I1007 23:44:54.443255 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-gtk4b" event={"ID":"d021d8ab-b7ae-4136-a946-2729bdd971b9","Type":"ContainerDied","Data":"4a2c2efd8c6101f6ff7df800e3a4a0a8b13b16ff72ae1ec43233a855fb874ea2"} Oct 07 23:44:55 crc kubenswrapper[4871]: I1007 23:44:55.848825 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-gtk4b" Oct 07 23:44:56 crc kubenswrapper[4871]: I1007 23:44:56.001562 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9rwp\" (UniqueName: \"kubernetes.io/projected/d021d8ab-b7ae-4136-a946-2729bdd971b9-kube-api-access-k9rwp\") pod \"d021d8ab-b7ae-4136-a946-2729bdd971b9\" (UID: \"d021d8ab-b7ae-4136-a946-2729bdd971b9\") " Oct 07 23:44:56 crc kubenswrapper[4871]: I1007 23:44:56.007912 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d021d8ab-b7ae-4136-a946-2729bdd971b9-kube-api-access-k9rwp" (OuterVolumeSpecName: "kube-api-access-k9rwp") pod "d021d8ab-b7ae-4136-a946-2729bdd971b9" (UID: "d021d8ab-b7ae-4136-a946-2729bdd971b9"). InnerVolumeSpecName "kube-api-access-k9rwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:44:56 crc kubenswrapper[4871]: I1007 23:44:56.106394 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9rwp\" (UniqueName: \"kubernetes.io/projected/d021d8ab-b7ae-4136-a946-2729bdd971b9-kube-api-access-k9rwp\") on node \"crc\" DevicePath \"\"" Oct 07 23:44:56 crc kubenswrapper[4871]: I1007 23:44:56.471366 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-gtk4b" event={"ID":"d021d8ab-b7ae-4136-a946-2729bdd971b9","Type":"ContainerDied","Data":"ac6423c8d68a70e2bae2ac41620abb5f2583f84cd4597c81d7e7020002556ad2"} Oct 07 23:44:56 crc kubenswrapper[4871]: I1007 23:44:56.471933 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac6423c8d68a70e2bae2ac41620abb5f2583f84cd4597c81d7e7020002556ad2" Oct 07 23:44:56 crc kubenswrapper[4871]: I1007 23:44:56.471473 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-gtk4b" Oct 07 23:44:58 crc kubenswrapper[4871]: I1007 23:44:58.982754 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:44:58 crc kubenswrapper[4871]: E1007 23:44:58.983572 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.163604 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl"] Oct 07 23:45:00 crc kubenswrapper[4871]: E1007 23:45:00.164168 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d021d8ab-b7ae-4136-a946-2729bdd971b9" containerName="mariadb-database-create" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.164189 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d021d8ab-b7ae-4136-a946-2729bdd971b9" containerName="mariadb-database-create" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.164486 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d021d8ab-b7ae-4136-a946-2729bdd971b9" containerName="mariadb-database-create" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.165333 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.167922 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.168493 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.171673 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl"] Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.289987 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e55c10e-3977-4d16-9ecf-769afad5ea95-secret-volume\") pod \"collect-profiles-29331345-4bbkl\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.290051 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e55c10e-3977-4d16-9ecf-769afad5ea95-config-volume\") pod \"collect-profiles-29331345-4bbkl\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.290078 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clsbw\" (UniqueName: \"kubernetes.io/projected/9e55c10e-3977-4d16-9ecf-769afad5ea95-kube-api-access-clsbw\") pod \"collect-profiles-29331345-4bbkl\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.391899 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e55c10e-3977-4d16-9ecf-769afad5ea95-secret-volume\") pod \"collect-profiles-29331345-4bbkl\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.392038 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e55c10e-3977-4d16-9ecf-769afad5ea95-config-volume\") pod \"collect-profiles-29331345-4bbkl\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.392097 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clsbw\" (UniqueName: \"kubernetes.io/projected/9e55c10e-3977-4d16-9ecf-769afad5ea95-kube-api-access-clsbw\") pod \"collect-profiles-29331345-4bbkl\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.393319 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e55c10e-3977-4d16-9ecf-769afad5ea95-config-volume\") pod \"collect-profiles-29331345-4bbkl\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.398304 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e55c10e-3977-4d16-9ecf-769afad5ea95-secret-volume\") pod \"collect-profiles-29331345-4bbkl\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.426694 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clsbw\" (UniqueName: \"kubernetes.io/projected/9e55c10e-3977-4d16-9ecf-769afad5ea95-kube-api-access-clsbw\") pod \"collect-profiles-29331345-4bbkl\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.503691 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:00 crc kubenswrapper[4871]: I1007 23:45:00.751758 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl"] Oct 07 23:45:01 crc kubenswrapper[4871]: I1007 23:45:01.524563 4871 generic.go:334] "Generic (PLEG): container finished" podID="9e55c10e-3977-4d16-9ecf-769afad5ea95" containerID="de57788daf996cb4b12b510173765a62c8ee548135e19b0f772efa3fb5283b40" exitCode=0 Oct 07 23:45:01 crc kubenswrapper[4871]: I1007 23:45:01.524764 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" event={"ID":"9e55c10e-3977-4d16-9ecf-769afad5ea95","Type":"ContainerDied","Data":"de57788daf996cb4b12b510173765a62c8ee548135e19b0f772efa3fb5283b40"} Oct 07 23:45:01 crc kubenswrapper[4871]: I1007 23:45:01.524967 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" event={"ID":"9e55c10e-3977-4d16-9ecf-769afad5ea95","Type":"ContainerStarted","Data":"4e65d337d7fea5c8111b4a2757f959480ef0e9b8e66d43b1a257d699bce8c54e"} Oct 07 23:45:02 crc kubenswrapper[4871]: I1007 23:45:02.621418 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-6084-account-create-jcwkq"] Oct 07 23:45:02 crc kubenswrapper[4871]: I1007 23:45:02.623135 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6084-account-create-jcwkq" Oct 07 23:45:02 crc kubenswrapper[4871]: I1007 23:45:02.625172 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 07 23:45:02 crc kubenswrapper[4871]: I1007 23:45:02.626968 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6084-account-create-jcwkq"] Oct 07 23:45:02 crc kubenswrapper[4871]: I1007 23:45:02.702297 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp8ws\" (UniqueName: \"kubernetes.io/projected/8d193011-4f44-4db3-837a-16116f2ab419-kube-api-access-mp8ws\") pod \"cinder-6084-account-create-jcwkq\" (UID: \"8d193011-4f44-4db3-837a-16116f2ab419\") " pod="openstack/cinder-6084-account-create-jcwkq" Oct 07 23:45:02 crc kubenswrapper[4871]: I1007 23:45:02.803766 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp8ws\" (UniqueName: \"kubernetes.io/projected/8d193011-4f44-4db3-837a-16116f2ab419-kube-api-access-mp8ws\") pod \"cinder-6084-account-create-jcwkq\" (UID: \"8d193011-4f44-4db3-837a-16116f2ab419\") " pod="openstack/cinder-6084-account-create-jcwkq" Oct 07 23:45:02 crc kubenswrapper[4871]: I1007 23:45:02.845373 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp8ws\" (UniqueName: \"kubernetes.io/projected/8d193011-4f44-4db3-837a-16116f2ab419-kube-api-access-mp8ws\") pod \"cinder-6084-account-create-jcwkq\" (UID: \"8d193011-4f44-4db3-837a-16116f2ab419\") " pod="openstack/cinder-6084-account-create-jcwkq" Oct 07 23:45:02 crc kubenswrapper[4871]: I1007 23:45:02.917239 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:02 crc kubenswrapper[4871]: I1007 23:45:02.957392 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6084-account-create-jcwkq" Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.006826 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e55c10e-3977-4d16-9ecf-769afad5ea95-config-volume\") pod \"9e55c10e-3977-4d16-9ecf-769afad5ea95\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.006910 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e55c10e-3977-4d16-9ecf-769afad5ea95-secret-volume\") pod \"9e55c10e-3977-4d16-9ecf-769afad5ea95\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.007076 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clsbw\" (UniqueName: \"kubernetes.io/projected/9e55c10e-3977-4d16-9ecf-769afad5ea95-kube-api-access-clsbw\") pod \"9e55c10e-3977-4d16-9ecf-769afad5ea95\" (UID: \"9e55c10e-3977-4d16-9ecf-769afad5ea95\") " Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.007679 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e55c10e-3977-4d16-9ecf-769afad5ea95-config-volume" (OuterVolumeSpecName: "config-volume") pod "9e55c10e-3977-4d16-9ecf-769afad5ea95" (UID: "9e55c10e-3977-4d16-9ecf-769afad5ea95"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.012297 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e55c10e-3977-4d16-9ecf-769afad5ea95-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9e55c10e-3977-4d16-9ecf-769afad5ea95" (UID: "9e55c10e-3977-4d16-9ecf-769afad5ea95"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.012690 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e55c10e-3977-4d16-9ecf-769afad5ea95-kube-api-access-clsbw" (OuterVolumeSpecName: "kube-api-access-clsbw") pod "9e55c10e-3977-4d16-9ecf-769afad5ea95" (UID: "9e55c10e-3977-4d16-9ecf-769afad5ea95"). InnerVolumeSpecName "kube-api-access-clsbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.109457 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clsbw\" (UniqueName: \"kubernetes.io/projected/9e55c10e-3977-4d16-9ecf-769afad5ea95-kube-api-access-clsbw\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.109765 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e55c10e-3977-4d16-9ecf-769afad5ea95-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.109774 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e55c10e-3977-4d16-9ecf-769afad5ea95-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.390937 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6084-account-create-jcwkq"] Oct 07 23:45:03 crc kubenswrapper[4871]: W1007 23:45:03.396226 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d193011_4f44_4db3_837a_16116f2ab419.slice/crio-fb3beaddb8b9c45a60c71876d84d058c79884aa45c3046f96c085c17e07a31be WatchSource:0}: Error finding container fb3beaddb8b9c45a60c71876d84d058c79884aa45c3046f96c085c17e07a31be: Status 404 returned error can't find the container with id fb3beaddb8b9c45a60c71876d84d058c79884aa45c3046f96c085c17e07a31be Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.545150 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6084-account-create-jcwkq" event={"ID":"8d193011-4f44-4db3-837a-16116f2ab419","Type":"ContainerStarted","Data":"fb3beaddb8b9c45a60c71876d84d058c79884aa45c3046f96c085c17e07a31be"} Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.546591 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" event={"ID":"9e55c10e-3977-4d16-9ecf-769afad5ea95","Type":"ContainerDied","Data":"4e65d337d7fea5c8111b4a2757f959480ef0e9b8e66d43b1a257d699bce8c54e"} Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.546622 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e65d337d7fea5c8111b4a2757f959480ef0e9b8e66d43b1a257d699bce8c54e" Oct 07 23:45:03 crc kubenswrapper[4871]: I1007 23:45:03.546679 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl" Oct 07 23:45:04 crc kubenswrapper[4871]: I1007 23:45:04.031337 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww"] Oct 07 23:45:04 crc kubenswrapper[4871]: I1007 23:45:04.042693 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331300-4llww"] Oct 07 23:45:04 crc kubenswrapper[4871]: I1007 23:45:04.561117 4871 generic.go:334] "Generic (PLEG): container finished" podID="8d193011-4f44-4db3-837a-16116f2ab419" containerID="d4d879ff4ca478c33d8fc835f8d12c1379f8c8ffc4d6bee5aeb7a0dec0959c6c" exitCode=0 Oct 07 23:45:04 crc kubenswrapper[4871]: I1007 23:45:04.561159 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6084-account-create-jcwkq" event={"ID":"8d193011-4f44-4db3-837a-16116f2ab419","Type":"ContainerDied","Data":"d4d879ff4ca478c33d8fc835f8d12c1379f8c8ffc4d6bee5aeb7a0dec0959c6c"} Oct 07 23:45:05 crc kubenswrapper[4871]: I1007 23:45:05.002484 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="034e49d6-9455-40d9-98fa-83d7b737a3c0" path="/var/lib/kubelet/pods/034e49d6-9455-40d9-98fa-83d7b737a3c0/volumes" Oct 07 23:45:06 crc kubenswrapper[4871]: I1007 23:45:06.014090 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6084-account-create-jcwkq" Oct 07 23:45:06 crc kubenswrapper[4871]: I1007 23:45:06.076099 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp8ws\" (UniqueName: \"kubernetes.io/projected/8d193011-4f44-4db3-837a-16116f2ab419-kube-api-access-mp8ws\") pod \"8d193011-4f44-4db3-837a-16116f2ab419\" (UID: \"8d193011-4f44-4db3-837a-16116f2ab419\") " Oct 07 23:45:06 crc kubenswrapper[4871]: I1007 23:45:06.084691 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d193011-4f44-4db3-837a-16116f2ab419-kube-api-access-mp8ws" (OuterVolumeSpecName: "kube-api-access-mp8ws") pod "8d193011-4f44-4db3-837a-16116f2ab419" (UID: "8d193011-4f44-4db3-837a-16116f2ab419"). InnerVolumeSpecName "kube-api-access-mp8ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:06 crc kubenswrapper[4871]: I1007 23:45:06.179159 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp8ws\" (UniqueName: \"kubernetes.io/projected/8d193011-4f44-4db3-837a-16116f2ab419-kube-api-access-mp8ws\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:06 crc kubenswrapper[4871]: I1007 23:45:06.580683 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6084-account-create-jcwkq" event={"ID":"8d193011-4f44-4db3-837a-16116f2ab419","Type":"ContainerDied","Data":"fb3beaddb8b9c45a60c71876d84d058c79884aa45c3046f96c085c17e07a31be"} Oct 07 23:45:06 crc kubenswrapper[4871]: I1007 23:45:06.581411 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb3beaddb8b9c45a60c71876d84d058c79884aa45c3046f96c085c17e07a31be" Oct 07 23:45:06 crc kubenswrapper[4871]: I1007 23:45:06.580834 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6084-account-create-jcwkq" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.871612 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-86mch"] Oct 07 23:45:07 crc kubenswrapper[4871]: E1007 23:45:07.872323 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e55c10e-3977-4d16-9ecf-769afad5ea95" containerName="collect-profiles" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.872347 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e55c10e-3977-4d16-9ecf-769afad5ea95" containerName="collect-profiles" Oct 07 23:45:07 crc kubenswrapper[4871]: E1007 23:45:07.872373 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d193011-4f44-4db3-837a-16116f2ab419" containerName="mariadb-account-create" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.872386 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d193011-4f44-4db3-837a-16116f2ab419" containerName="mariadb-account-create" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.872737 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e55c10e-3977-4d16-9ecf-769afad5ea95" containerName="collect-profiles" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.872827 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d193011-4f44-4db3-837a-16116f2ab419" containerName="mariadb-account-create" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.873925 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.876073 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.877271 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rgv49" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.879261 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.886866 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-86mch"] Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.911843 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg4sb\" (UniqueName: \"kubernetes.io/projected/3ede6bf1-ea5c-4298-9c61-5330dfc70805-kube-api-access-hg4sb\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.912036 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ede6bf1-ea5c-4298-9c61-5330dfc70805-etc-machine-id\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.912131 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-db-sync-config-data\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.912167 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-config-data\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.912224 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-scripts\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:07 crc kubenswrapper[4871]: I1007 23:45:07.912339 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-combined-ca-bundle\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.014548 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-combined-ca-bundle\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.014650 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg4sb\" (UniqueName: \"kubernetes.io/projected/3ede6bf1-ea5c-4298-9c61-5330dfc70805-kube-api-access-hg4sb\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.014768 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ede6bf1-ea5c-4298-9c61-5330dfc70805-etc-machine-id\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.014879 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-db-sync-config-data\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.014910 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-config-data\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.014971 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-scripts\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.015077 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ede6bf1-ea5c-4298-9c61-5330dfc70805-etc-machine-id\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.019696 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-db-sync-config-data\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.019726 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-combined-ca-bundle\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.023575 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-scripts\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.025422 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-config-data\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.042961 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg4sb\" (UniqueName: \"kubernetes.io/projected/3ede6bf1-ea5c-4298-9c61-5330dfc70805-kube-api-access-hg4sb\") pod \"cinder-db-sync-86mch\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.207705 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:08 crc kubenswrapper[4871]: I1007 23:45:08.690844 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-86mch"] Oct 07 23:45:09 crc kubenswrapper[4871]: I1007 23:45:09.618536 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-86mch" event={"ID":"3ede6bf1-ea5c-4298-9c61-5330dfc70805","Type":"ContainerStarted","Data":"c3c1dc460812303077ce228afc16b680d56f90a68c150c03cc56a6b658a0096a"} Oct 07 23:45:09 crc kubenswrapper[4871]: I1007 23:45:09.618968 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-86mch" event={"ID":"3ede6bf1-ea5c-4298-9c61-5330dfc70805","Type":"ContainerStarted","Data":"6e3dd363a356df7fde84bee01277e544a8437e2eacb4b969ab755dd2a9828f17"} Oct 07 23:45:09 crc kubenswrapper[4871]: I1007 23:45:09.643859 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-86mch" podStartSLOduration=2.643837653 podStartE2EDuration="2.643837653s" podCreationTimestamp="2025-10-07 23:45:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:09.63728731 +0000 UTC m=+5783.439985383" watchObservedRunningTime="2025-10-07 23:45:09.643837653 +0000 UTC m=+5783.446535736" Oct 07 23:45:10 crc kubenswrapper[4871]: I1007 23:45:10.982346 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:45:10 crc kubenswrapper[4871]: E1007 23:45:10.982880 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:45:12 crc kubenswrapper[4871]: I1007 23:45:12.664518 4871 generic.go:334] "Generic (PLEG): container finished" podID="3ede6bf1-ea5c-4298-9c61-5330dfc70805" containerID="c3c1dc460812303077ce228afc16b680d56f90a68c150c03cc56a6b658a0096a" exitCode=0 Oct 07 23:45:12 crc kubenswrapper[4871]: I1007 23:45:12.664785 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-86mch" event={"ID":"3ede6bf1-ea5c-4298-9c61-5330dfc70805","Type":"ContainerDied","Data":"c3c1dc460812303077ce228afc16b680d56f90a68c150c03cc56a6b658a0096a"} Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.064120 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.149000 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-combined-ca-bundle\") pod \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.149085 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ede6bf1-ea5c-4298-9c61-5330dfc70805-etc-machine-id\") pod \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.149133 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-db-sync-config-data\") pod \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.149182 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg4sb\" (UniqueName: \"kubernetes.io/projected/3ede6bf1-ea5c-4298-9c61-5330dfc70805-kube-api-access-hg4sb\") pod \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.149198 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ede6bf1-ea5c-4298-9c61-5330dfc70805-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3ede6bf1-ea5c-4298-9c61-5330dfc70805" (UID: "3ede6bf1-ea5c-4298-9c61-5330dfc70805"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.149232 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-scripts\") pod \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.149292 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-config-data\") pod \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\" (UID: \"3ede6bf1-ea5c-4298-9c61-5330dfc70805\") " Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.149672 4871 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3ede6bf1-ea5c-4298-9c61-5330dfc70805-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.154280 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3ede6bf1-ea5c-4298-9c61-5330dfc70805" (UID: "3ede6bf1-ea5c-4298-9c61-5330dfc70805"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.155425 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-scripts" (OuterVolumeSpecName: "scripts") pod "3ede6bf1-ea5c-4298-9c61-5330dfc70805" (UID: "3ede6bf1-ea5c-4298-9c61-5330dfc70805"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.159960 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ede6bf1-ea5c-4298-9c61-5330dfc70805-kube-api-access-hg4sb" (OuterVolumeSpecName: "kube-api-access-hg4sb") pod "3ede6bf1-ea5c-4298-9c61-5330dfc70805" (UID: "3ede6bf1-ea5c-4298-9c61-5330dfc70805"). InnerVolumeSpecName "kube-api-access-hg4sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.178122 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ede6bf1-ea5c-4298-9c61-5330dfc70805" (UID: "3ede6bf1-ea5c-4298-9c61-5330dfc70805"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.204113 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-config-data" (OuterVolumeSpecName: "config-data") pod "3ede6bf1-ea5c-4298-9c61-5330dfc70805" (UID: "3ede6bf1-ea5c-4298-9c61-5330dfc70805"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.253512 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.253558 4871 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.253570 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg4sb\" (UniqueName: \"kubernetes.io/projected/3ede6bf1-ea5c-4298-9c61-5330dfc70805-kube-api-access-hg4sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.253603 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.253614 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ede6bf1-ea5c-4298-9c61-5330dfc70805-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.732763 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-86mch" event={"ID":"3ede6bf1-ea5c-4298-9c61-5330dfc70805","Type":"ContainerDied","Data":"6e3dd363a356df7fde84bee01277e544a8437e2eacb4b969ab755dd2a9828f17"} Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.732819 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e3dd363a356df7fde84bee01277e544a8437e2eacb4b969ab755dd2a9828f17" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.732900 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-86mch" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.998929 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bd78b65c9-6mlmx"] Oct 07 23:45:14 crc kubenswrapper[4871]: E1007 23:45:14.999775 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ede6bf1-ea5c-4298-9c61-5330dfc70805" containerName="cinder-db-sync" Oct 07 23:45:14 crc kubenswrapper[4871]: I1007 23:45:14.999891 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ede6bf1-ea5c-4298-9c61-5330dfc70805" containerName="cinder-db-sync" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.000145 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ede6bf1-ea5c-4298-9c61-5330dfc70805" containerName="cinder-db-sync" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.001160 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.013786 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd78b65c9-6mlmx"] Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.069989 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6spq\" (UniqueName: \"kubernetes.io/projected/15b9cadf-0dd4-4f1d-a372-84dee39459cd-kube-api-access-z6spq\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.073625 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.073768 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-dns-svc\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.073854 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.073951 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-config\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.133268 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.134675 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.138067 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rgv49" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.138304 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.138421 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.138508 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.150944 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175295 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175344 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af866d9d-14f4-4e78-82f9-01102c64aea5-logs\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175373 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6spq\" (UniqueName: \"kubernetes.io/projected/15b9cadf-0dd4-4f1d-a372-84dee39459cd-kube-api-access-z6spq\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175394 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data-custom\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175449 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpf65\" (UniqueName: \"kubernetes.io/projected/af866d9d-14f4-4e78-82f9-01102c64aea5-kube-api-access-wpf65\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175465 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175480 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-scripts\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175498 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175526 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-dns-svc\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175540 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175560 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-config\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.175609 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af866d9d-14f4-4e78-82f9-01102c64aea5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.177432 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.177467 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-config\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.177482 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-dns-svc\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.177589 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.212966 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6spq\" (UniqueName: \"kubernetes.io/projected/15b9cadf-0dd4-4f1d-a372-84dee39459cd-kube-api-access-z6spq\") pod \"dnsmasq-dns-6bd78b65c9-6mlmx\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.278359 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.278403 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af866d9d-14f4-4e78-82f9-01102c64aea5-logs\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.278448 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data-custom\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.278503 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpf65\" (UniqueName: \"kubernetes.io/projected/af866d9d-14f4-4e78-82f9-01102c64aea5-kube-api-access-wpf65\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.278525 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.278554 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-scripts\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.278603 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af866d9d-14f4-4e78-82f9-01102c64aea5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.278691 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af866d9d-14f4-4e78-82f9-01102c64aea5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.279359 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af866d9d-14f4-4e78-82f9-01102c64aea5-logs\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.286149 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.286350 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data-custom\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.290015 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-scripts\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.291531 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.294998 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpf65\" (UniqueName: \"kubernetes.io/projected/af866d9d-14f4-4e78-82f9-01102c64aea5-kube-api-access-wpf65\") pod \"cinder-api-0\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.329089 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.461978 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.822440 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd78b65c9-6mlmx"] Oct 07 23:45:15 crc kubenswrapper[4871]: W1007 23:45:15.830118 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15b9cadf_0dd4_4f1d_a372_84dee39459cd.slice/crio-17997db74971439eb59c77b73d35454f8a2153e3ad89b0d999319ab9976b62a4 WatchSource:0}: Error finding container 17997db74971439eb59c77b73d35454f8a2153e3ad89b0d999319ab9976b62a4: Status 404 returned error can't find the container with id 17997db74971439eb59c77b73d35454f8a2153e3ad89b0d999319ab9976b62a4 Oct 07 23:45:15 crc kubenswrapper[4871]: I1007 23:45:15.951331 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 23:45:15 crc kubenswrapper[4871]: W1007 23:45:15.964750 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf866d9d_14f4_4e78_82f9_01102c64aea5.slice/crio-520a6f412c4bba09a2941757238cd87336cf860ec61ec04efe5c630909b7c189 WatchSource:0}: Error finding container 520a6f412c4bba09a2941757238cd87336cf860ec61ec04efe5c630909b7c189: Status 404 returned error can't find the container with id 520a6f412c4bba09a2941757238cd87336cf860ec61ec04efe5c630909b7c189 Oct 07 23:45:16 crc kubenswrapper[4871]: I1007 23:45:16.758451 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af866d9d-14f4-4e78-82f9-01102c64aea5","Type":"ContainerStarted","Data":"80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67"} Oct 07 23:45:16 crc kubenswrapper[4871]: I1007 23:45:16.760571 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af866d9d-14f4-4e78-82f9-01102c64aea5","Type":"ContainerStarted","Data":"520a6f412c4bba09a2941757238cd87336cf860ec61ec04efe5c630909b7c189"} Oct 07 23:45:16 crc kubenswrapper[4871]: I1007 23:45:16.761537 4871 generic.go:334] "Generic (PLEG): container finished" podID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" containerID="74e6dd2b302ee5246afaf7bc49ad6c5f82be1a5b5b12c8bfa5d797a856f9c1c0" exitCode=0 Oct 07 23:45:16 crc kubenswrapper[4871]: I1007 23:45:16.761646 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" event={"ID":"15b9cadf-0dd4-4f1d-a372-84dee39459cd","Type":"ContainerDied","Data":"74e6dd2b302ee5246afaf7bc49ad6c5f82be1a5b5b12c8bfa5d797a856f9c1c0"} Oct 07 23:45:16 crc kubenswrapper[4871]: I1007 23:45:16.761699 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" event={"ID":"15b9cadf-0dd4-4f1d-a372-84dee39459cd","Type":"ContainerStarted","Data":"17997db74971439eb59c77b73d35454f8a2153e3ad89b0d999319ab9976b62a4"} Oct 07 23:45:17 crc kubenswrapper[4871]: I1007 23:45:17.770291 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" event={"ID":"15b9cadf-0dd4-4f1d-a372-84dee39459cd","Type":"ContainerStarted","Data":"df1d813d2ce9d4336674ddcd51960dad2c4c71b3a17fe0f0baef521b98e1edbd"} Oct 07 23:45:17 crc kubenswrapper[4871]: I1007 23:45:17.771683 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:17 crc kubenswrapper[4871]: I1007 23:45:17.773557 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af866d9d-14f4-4e78-82f9-01102c64aea5","Type":"ContainerStarted","Data":"d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c"} Oct 07 23:45:17 crc kubenswrapper[4871]: I1007 23:45:17.773671 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 07 23:45:17 crc kubenswrapper[4871]: I1007 23:45:17.794741 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" podStartSLOduration=3.794720201 podStartE2EDuration="3.794720201s" podCreationTimestamp="2025-10-07 23:45:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:17.789424581 +0000 UTC m=+5791.592122654" watchObservedRunningTime="2025-10-07 23:45:17.794720201 +0000 UTC m=+5791.597418274" Oct 07 23:45:17 crc kubenswrapper[4871]: I1007 23:45:17.815689 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.815668234 podStartE2EDuration="2.815668234s" podCreationTimestamp="2025-10-07 23:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:17.806708007 +0000 UTC m=+5791.609406080" watchObservedRunningTime="2025-10-07 23:45:17.815668234 +0000 UTC m=+5791.618366317" Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.915113 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.915630 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="a69f3ff4-898a-4df6-9734-334249368868" containerName="nova-cell0-conductor-conductor" containerID="cri-o://7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8" gracePeriod=30 Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.944694 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.944968 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-log" containerID="cri-o://38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558" gracePeriod=30 Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.945276 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-metadata" containerID="cri-o://f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67" gracePeriod=30 Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.954384 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.954645 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5e592305-57bf-441b-926e-0217c88ab964" containerName="nova-scheduler-scheduler" containerID="cri-o://ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b" gracePeriod=30 Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.962077 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.962327 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-log" containerID="cri-o://8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1" gracePeriod=30 Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.962368 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-api" containerID="cri-o://c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1" gracePeriod=30 Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.972629 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 23:45:18 crc kubenswrapper[4871]: I1007 23:45:18.972873 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="4aefc2ec-aa7f-4782-8993-f48aba3434d3" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://5af235543317d416b3ea0b9fb8edceb8e1d2aca6a85e1c4f511e1521b38236f9" gracePeriod=30 Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.794644 4871 generic.go:334] "Generic (PLEG): container finished" podID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerID="38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558" exitCode=143 Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.795071 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f23c5ed1-6597-4d11-9270-6c0428a37d42","Type":"ContainerDied","Data":"38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558"} Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.797614 4871 generic.go:334] "Generic (PLEG): container finished" podID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerID="8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1" exitCode=143 Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.797691 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c","Type":"ContainerDied","Data":"8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1"} Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.799263 4871 generic.go:334] "Generic (PLEG): container finished" podID="4aefc2ec-aa7f-4782-8993-f48aba3434d3" containerID="5af235543317d416b3ea0b9fb8edceb8e1d2aca6a85e1c4f511e1521b38236f9" exitCode=0 Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.799317 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4aefc2ec-aa7f-4782-8993-f48aba3434d3","Type":"ContainerDied","Data":"5af235543317d416b3ea0b9fb8edceb8e1d2aca6a85e1c4f511e1521b38236f9"} Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.799336 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4aefc2ec-aa7f-4782-8993-f48aba3434d3","Type":"ContainerDied","Data":"e1fdfbf3f8f0728600d38ba684bc42966ef206870eb0c114ce41964979d5ae19"} Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.799368 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1fdfbf3f8f0728600d38ba684bc42966ef206870eb0c114ce41964979d5ae19" Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.802089 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.859166 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-combined-ca-bundle\") pod \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.859295 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t77t9\" (UniqueName: \"kubernetes.io/projected/4aefc2ec-aa7f-4782-8993-f48aba3434d3-kube-api-access-t77t9\") pod \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.859376 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-config-data\") pod \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\" (UID: \"4aefc2ec-aa7f-4782-8993-f48aba3434d3\") " Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.869690 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aefc2ec-aa7f-4782-8993-f48aba3434d3-kube-api-access-t77t9" (OuterVolumeSpecName: "kube-api-access-t77t9") pod "4aefc2ec-aa7f-4782-8993-f48aba3434d3" (UID: "4aefc2ec-aa7f-4782-8993-f48aba3434d3"). InnerVolumeSpecName "kube-api-access-t77t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.900030 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4aefc2ec-aa7f-4782-8993-f48aba3434d3" (UID: "4aefc2ec-aa7f-4782-8993-f48aba3434d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.903441 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-config-data" (OuterVolumeSpecName: "config-data") pod "4aefc2ec-aa7f-4782-8993-f48aba3434d3" (UID: "4aefc2ec-aa7f-4782-8993-f48aba3434d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.961015 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.961046 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aefc2ec-aa7f-4782-8993-f48aba3434d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:19 crc kubenswrapper[4871]: I1007 23:45:19.961057 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t77t9\" (UniqueName: \"kubernetes.io/projected/4aefc2ec-aa7f-4782-8993-f48aba3434d3-kube-api-access-t77t9\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.809476 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.841657 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.850783 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.868605 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 23:45:20 crc kubenswrapper[4871]: E1007 23:45:20.869357 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aefc2ec-aa7f-4782-8993-f48aba3434d3" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.869393 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aefc2ec-aa7f-4782-8993-f48aba3434d3" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.869639 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aefc2ec-aa7f-4782-8993-f48aba3434d3" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.870338 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.872475 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.877515 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ba2fb5-0f9f-41fa-816b-1db49317eb3e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.877616 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ba2fb5-0f9f-41fa-816b-1db49317eb3e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.877986 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6z7m\" (UniqueName: \"kubernetes.io/projected/c5ba2fb5-0f9f-41fa-816b-1db49317eb3e-kube-api-access-z6z7m\") pod \"nova-cell1-novncproxy-0\" (UID: \"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.898373 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.979122 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6z7m\" (UniqueName: \"kubernetes.io/projected/c5ba2fb5-0f9f-41fa-816b-1db49317eb3e-kube-api-access-z6z7m\") pod \"nova-cell1-novncproxy-0\" (UID: \"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.979193 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ba2fb5-0f9f-41fa-816b-1db49317eb3e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.979242 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ba2fb5-0f9f-41fa-816b-1db49317eb3e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.983325 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ba2fb5-0f9f-41fa-816b-1db49317eb3e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.983479 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ba2fb5-0f9f-41fa-816b-1db49317eb3e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.992877 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aefc2ec-aa7f-4782-8993-f48aba3434d3" path="/var/lib/kubelet/pods/4aefc2ec-aa7f-4782-8993-f48aba3434d3/volumes" Oct 07 23:45:20 crc kubenswrapper[4871]: I1007 23:45:20.995224 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6z7m\" (UniqueName: \"kubernetes.io/projected/c5ba2fb5-0f9f-41fa-816b-1db49317eb3e-kube-api-access-z6z7m\") pod \"nova-cell1-novncproxy-0\" (UID: \"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:21 crc kubenswrapper[4871]: I1007 23:45:21.193370 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:21 crc kubenswrapper[4871]: E1007 23:45:21.407604 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 23:45:21 crc kubenswrapper[4871]: E1007 23:45:21.410060 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 23:45:21 crc kubenswrapper[4871]: E1007 23:45:21.411557 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 23:45:21 crc kubenswrapper[4871]: E1007 23:45:21.411661 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="5e592305-57bf-441b-926e-0217c88ab964" containerName="nova-scheduler-scheduler" Oct 07 23:45:21 crc kubenswrapper[4871]: I1007 23:45:21.696403 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 23:45:21 crc kubenswrapper[4871]: I1007 23:45:21.842340 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e","Type":"ContainerStarted","Data":"ab21076d6c7f1dea3885867645d89831b754d00caac46a68d275ad7139453bcc"} Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.098393 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": read tcp 10.217.0.2:56012->10.217.1.69:8775: read: connection reset by peer" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.098403 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": read tcp 10.217.0.2:56018->10.217.1.69:8775: read: connection reset by peer" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.172133 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.172387 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="88b2fbdc-ee2b-4803-be96-c47713900f08" containerName="nova-cell1-conductor-conductor" containerID="cri-o://b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a" gracePeriod=30 Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.656293 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.663991 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.671455 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718339 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf65k\" (UniqueName: \"kubernetes.io/projected/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-kube-api-access-zf65k\") pod \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718437 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-config-data\") pod \"f23c5ed1-6597-4d11-9270-6c0428a37d42\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718502 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-combined-ca-bundle\") pod \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718561 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-logs\") pod \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718601 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-config-data\") pod \"a69f3ff4-898a-4df6-9734-334249368868\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718651 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvkqh\" (UniqueName: \"kubernetes.io/projected/f23c5ed1-6597-4d11-9270-6c0428a37d42-kube-api-access-gvkqh\") pod \"f23c5ed1-6597-4d11-9270-6c0428a37d42\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718719 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f23c5ed1-6597-4d11-9270-6c0428a37d42-logs\") pod \"f23c5ed1-6597-4d11-9270-6c0428a37d42\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718752 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-combined-ca-bundle\") pod \"f23c5ed1-6597-4d11-9270-6c0428a37d42\" (UID: \"f23c5ed1-6597-4d11-9270-6c0428a37d42\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718846 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-combined-ca-bundle\") pod \"a69f3ff4-898a-4df6-9734-334249368868\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718874 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-config-data\") pod \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\" (UID: \"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.718915 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljw74\" (UniqueName: \"kubernetes.io/projected/a69f3ff4-898a-4df6-9734-334249368868-kube-api-access-ljw74\") pod \"a69f3ff4-898a-4df6-9734-334249368868\" (UID: \"a69f3ff4-898a-4df6-9734-334249368868\") " Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.734497 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a69f3ff4-898a-4df6-9734-334249368868-kube-api-access-ljw74" (OuterVolumeSpecName: "kube-api-access-ljw74") pod "a69f3ff4-898a-4df6-9734-334249368868" (UID: "a69f3ff4-898a-4df6-9734-334249368868"). InnerVolumeSpecName "kube-api-access-ljw74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.736892 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-logs" (OuterVolumeSpecName: "logs") pod "5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" (UID: "5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.738584 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f23c5ed1-6597-4d11-9270-6c0428a37d42-logs" (OuterVolumeSpecName: "logs") pod "f23c5ed1-6597-4d11-9270-6c0428a37d42" (UID: "f23c5ed1-6597-4d11-9270-6c0428a37d42"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.740269 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-kube-api-access-zf65k" (OuterVolumeSpecName: "kube-api-access-zf65k") pod "5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" (UID: "5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c"). InnerVolumeSpecName "kube-api-access-zf65k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.750651 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f23c5ed1-6597-4d11-9270-6c0428a37d42-kube-api-access-gvkqh" (OuterVolumeSpecName: "kube-api-access-gvkqh") pod "f23c5ed1-6597-4d11-9270-6c0428a37d42" (UID: "f23c5ed1-6597-4d11-9270-6c0428a37d42"). InnerVolumeSpecName "kube-api-access-gvkqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.767408 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-config-data" (OuterVolumeSpecName: "config-data") pod "a69f3ff4-898a-4df6-9734-334249368868" (UID: "a69f3ff4-898a-4df6-9734-334249368868"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.771910 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f23c5ed1-6597-4d11-9270-6c0428a37d42" (UID: "f23c5ed1-6597-4d11-9270-6c0428a37d42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.780322 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-config-data" (OuterVolumeSpecName: "config-data") pod "f23c5ed1-6597-4d11-9270-6c0428a37d42" (UID: "f23c5ed1-6597-4d11-9270-6c0428a37d42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.790151 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a69f3ff4-898a-4df6-9734-334249368868" (UID: "a69f3ff4-898a-4df6-9734-334249368868"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.792944 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" (UID: "5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.794457 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-config-data" (OuterVolumeSpecName: "config-data") pod "5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" (UID: "5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820497 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820532 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820543 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljw74\" (UniqueName: \"kubernetes.io/projected/a69f3ff4-898a-4df6-9734-334249368868-kube-api-access-ljw74\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820554 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf65k\" (UniqueName: \"kubernetes.io/projected/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-kube-api-access-zf65k\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820563 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820571 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820580 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820590 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a69f3ff4-898a-4df6-9734-334249368868-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820598 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvkqh\" (UniqueName: \"kubernetes.io/projected/f23c5ed1-6597-4d11-9270-6c0428a37d42-kube-api-access-gvkqh\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820606 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f23c5ed1-6597-4d11-9270-6c0428a37d42-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.820614 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23c5ed1-6597-4d11-9270-6c0428a37d42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.862373 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c5ba2fb5-0f9f-41fa-816b-1db49317eb3e","Type":"ContainerStarted","Data":"b4f8637d471ef96157740c65e19f5551f9fea02d5cd9d03e9d580ac9ee6f36a1"} Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.864494 4871 generic.go:334] "Generic (PLEG): container finished" podID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerID="c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1" exitCode=0 Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.864537 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c","Type":"ContainerDied","Data":"c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1"} Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.864555 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c","Type":"ContainerDied","Data":"3e7f63389a5f6c0c06f8ed5ea9a4f5d30e654898d9eebf7428f4c5851d4ff5da"} Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.864569 4871 scope.go:117] "RemoveContainer" containerID="c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.864626 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.880592 4871 generic.go:334] "Generic (PLEG): container finished" podID="a69f3ff4-898a-4df6-9734-334249368868" containerID="7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8" exitCode=0 Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.880642 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a69f3ff4-898a-4df6-9734-334249368868","Type":"ContainerDied","Data":"7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8"} Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.880666 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a69f3ff4-898a-4df6-9734-334249368868","Type":"ContainerDied","Data":"a7dfb752a70e6c420990ffe862c893a72e7c2b0016a74bab466cb62bf8a5d21b"} Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.880725 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.882761 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.88273649 podStartE2EDuration="2.88273649s" podCreationTimestamp="2025-10-07 23:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:22.879146045 +0000 UTC m=+5796.681844118" watchObservedRunningTime="2025-10-07 23:45:22.88273649 +0000 UTC m=+5796.685434573" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.886346 4871 generic.go:334] "Generic (PLEG): container finished" podID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerID="f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67" exitCode=0 Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.886385 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f23c5ed1-6597-4d11-9270-6c0428a37d42","Type":"ContainerDied","Data":"f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67"} Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.886423 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f23c5ed1-6597-4d11-9270-6c0428a37d42","Type":"ContainerDied","Data":"5ed53e44433f6372af212c04d09fbbc0ae3d8e2bbd9d7a301350f160836609c0"} Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.886532 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.910713 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.925027 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.928395 4871 scope.go:117] "RemoveContainer" containerID="8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.937080 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.950147 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.957749 4871 scope.go:117] "RemoveContainer" containerID="c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1" Oct 07 23:45:22 crc kubenswrapper[4871]: E1007 23:45:22.958145 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1\": container with ID starting with c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1 not found: ID does not exist" containerID="c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.958166 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1"} err="failed to get container status \"c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1\": rpc error: code = NotFound desc = could not find container \"c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1\": container with ID starting with c6866941a9a53b110b35cba0756f90339cc77ef0e81f290f22b21692f0e90ab1 not found: ID does not exist" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.958183 4871 scope.go:117] "RemoveContainer" containerID="8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1" Oct 07 23:45:22 crc kubenswrapper[4871]: E1007 23:45:22.958353 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1\": container with ID starting with 8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1 not found: ID does not exist" containerID="8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.958368 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1"} err="failed to get container status \"8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1\": rpc error: code = NotFound desc = could not find container \"8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1\": container with ID starting with 8a1fab924cf3bc9b13517ab25e02c5cee6e1c126860e0f2e59bc33ff49a949f1 not found: ID does not exist" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.958389 4871 scope.go:117] "RemoveContainer" containerID="7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.964941 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 23:45:22 crc kubenswrapper[4871]: E1007 23:45:22.965338 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69f3ff4-898a-4df6-9734-334249368868" containerName="nova-cell0-conductor-conductor" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965349 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69f3ff4-898a-4df6-9734-334249368868" containerName="nova-cell0-conductor-conductor" Oct 07 23:45:22 crc kubenswrapper[4871]: E1007 23:45:22.965364 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-log" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965371 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-log" Oct 07 23:45:22 crc kubenswrapper[4871]: E1007 23:45:22.965387 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-log" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965393 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-log" Oct 07 23:45:22 crc kubenswrapper[4871]: E1007 23:45:22.965414 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-metadata" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965420 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-metadata" Oct 07 23:45:22 crc kubenswrapper[4871]: E1007 23:45:22.965432 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-api" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965438 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-api" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965616 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-api" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965634 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-metadata" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965646 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" containerName="nova-api-log" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965656 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" containerName="nova-metadata-log" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.965666 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69f3ff4-898a-4df6-9734-334249368868" containerName="nova-cell0-conductor-conductor" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.966610 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.978020 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.979222 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.984269 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.984914 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.985713 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:45:22 crc kubenswrapper[4871]: E1007 23:45:22.985922 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.989251 4871 scope.go:117] "RemoveContainer" containerID="7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8" Oct 07 23:45:22 crc kubenswrapper[4871]: E1007 23:45:22.989623 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8\": container with ID starting with 7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8 not found: ID does not exist" containerID="7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.989645 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8"} err="failed to get container status \"7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8\": rpc error: code = NotFound desc = could not find container \"7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8\": container with ID starting with 7788e79d7c9b936abcb1fb20a719cb48d700525850eff2a46debe35d9703afa8 not found: ID does not exist" Oct 07 23:45:22 crc kubenswrapper[4871]: I1007 23:45:22.989662 4871 scope.go:117] "RemoveContainer" containerID="f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.002109 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c" path="/var/lib/kubelet/pods/5e5ea6bf-7e9c-4dd2-9c9c-6c98b9e7280c/volumes" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.002667 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a69f3ff4-898a-4df6-9734-334249368868" path="/var/lib/kubelet/pods/a69f3ff4-898a-4df6-9734-334249368868/volumes" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.005260 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.005288 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.013576 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.019855 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.024242 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.024312 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r92x7\" (UniqueName: \"kubernetes.io/projected/426135f0-8fe9-40e8-9a1d-f95c000de11a-kube-api-access-r92x7\") pod \"nova-cell0-conductor-0\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.024356 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82n4l\" (UniqueName: \"kubernetes.io/projected/a313c68a-aeef-4003-a77d-97fb7ba6b22c-kube-api-access-82n4l\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.024440 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.024496 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a313c68a-aeef-4003-a77d-97fb7ba6b22c-logs\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.024645 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-config-data\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.024705 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.025917 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.027532 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.032953 4871 scope.go:117] "RemoveContainer" containerID="38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.033062 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.033391 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.061470 4871 scope.go:117] "RemoveContainer" containerID="f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67" Oct 07 23:45:23 crc kubenswrapper[4871]: E1007 23:45:23.061920 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67\": container with ID starting with f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67 not found: ID does not exist" containerID="f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.061960 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67"} err="failed to get container status \"f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67\": rpc error: code = NotFound desc = could not find container \"f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67\": container with ID starting with f16f161442763e0067e003684a826e8487956d54ad3269738403db47dd933f67 not found: ID does not exist" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.061984 4871 scope.go:117] "RemoveContainer" containerID="38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558" Oct 07 23:45:23 crc kubenswrapper[4871]: E1007 23:45:23.062280 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558\": container with ID starting with 38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558 not found: ID does not exist" containerID="38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.062329 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558"} err="failed to get container status \"38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558\": rpc error: code = NotFound desc = could not find container \"38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558\": container with ID starting with 38d2e30b872d1a14bdd06dd74a2e145a9390ae42e17736b9e32c1f6fa9986558 not found: ID does not exist" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.126704 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-config-data\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127231 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127369 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d2x8\" (UniqueName: \"kubernetes.io/projected/7fe15a5e-9ee6-47ba-a450-044b62740488-kube-api-access-7d2x8\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127432 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127489 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127571 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r92x7\" (UniqueName: \"kubernetes.io/projected/426135f0-8fe9-40e8-9a1d-f95c000de11a-kube-api-access-r92x7\") pod \"nova-cell0-conductor-0\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127633 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82n4l\" (UniqueName: \"kubernetes.io/projected/a313c68a-aeef-4003-a77d-97fb7ba6b22c-kube-api-access-82n4l\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127711 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-config-data\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127795 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127883 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe15a5e-9ee6-47ba-a450-044b62740488-logs\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.127981 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a313c68a-aeef-4003-a77d-97fb7ba6b22c-logs\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.128765 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a313c68a-aeef-4003-a77d-97fb7ba6b22c-logs\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.130458 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.130516 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.131642 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-config-data\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.136033 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.145615 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82n4l\" (UniqueName: \"kubernetes.io/projected/a313c68a-aeef-4003-a77d-97fb7ba6b22c-kube-api-access-82n4l\") pod \"nova-api-0\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.149679 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r92x7\" (UniqueName: \"kubernetes.io/projected/426135f0-8fe9-40e8-9a1d-f95c000de11a-kube-api-access-r92x7\") pod \"nova-cell0-conductor-0\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.228951 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d2x8\" (UniqueName: \"kubernetes.io/projected/7fe15a5e-9ee6-47ba-a450-044b62740488-kube-api-access-7d2x8\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.228998 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.229041 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-config-data\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.229070 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe15a5e-9ee6-47ba-a450-044b62740488-logs\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.230763 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe15a5e-9ee6-47ba-a450-044b62740488-logs\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.243865 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.243865 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-config-data\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.247290 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d2x8\" (UniqueName: \"kubernetes.io/projected/7fe15a5e-9ee6-47ba-a450-044b62740488-kube-api-access-7d2x8\") pod \"nova-metadata-0\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.298012 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.306439 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.357851 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.829106 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 23:45:23 crc kubenswrapper[4871]: W1007 23:45:23.841059 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda313c68a_aeef_4003_a77d_97fb7ba6b22c.slice/crio-34ab9a938bd7ce476e94621b1fe895d0f1bb56afc324e758fdb676767eb61c77 WatchSource:0}: Error finding container 34ab9a938bd7ce476e94621b1fe895d0f1bb56afc324e758fdb676767eb61c77: Status 404 returned error can't find the container with id 34ab9a938bd7ce476e94621b1fe895d0f1bb56afc324e758fdb676767eb61c77 Oct 07 23:45:23 crc kubenswrapper[4871]: W1007 23:45:23.849302 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod426135f0_8fe9_40e8_9a1d_f95c000de11a.slice/crio-9da9e043413052afc73e2a81275d4e002e5fd4e60a5dd9833aa841028af4e81f WatchSource:0}: Error finding container 9da9e043413052afc73e2a81275d4e002e5fd4e60a5dd9833aa841028af4e81f: Status 404 returned error can't find the container with id 9da9e043413052afc73e2a81275d4e002e5fd4e60a5dd9833aa841028af4e81f Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.851810 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.908251 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"426135f0-8fe9-40e8-9a1d-f95c000de11a","Type":"ContainerStarted","Data":"9da9e043413052afc73e2a81275d4e002e5fd4e60a5dd9833aa841028af4e81f"} Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.926462 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a313c68a-aeef-4003-a77d-97fb7ba6b22c","Type":"ContainerStarted","Data":"34ab9a938bd7ce476e94621b1fe895d0f1bb56afc324e758fdb676767eb61c77"} Oct 07 23:45:23 crc kubenswrapper[4871]: I1007 23:45:23.942189 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.764557 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="4aefc2ec-aa7f-4782-8993-f48aba3434d3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.58:6080/vnc_lite.html\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.935818 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fe15a5e-9ee6-47ba-a450-044b62740488","Type":"ContainerStarted","Data":"2a76e013141cc21f95e127a57c2b86803526962045fe5a12a9ad4f1f8656b434"} Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.935875 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fe15a5e-9ee6-47ba-a450-044b62740488","Type":"ContainerStarted","Data":"6a75584a8f6ac34bebf2de3afad9e86ebd85f9ba6fb3d711a0941a19bd274f39"} Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.935888 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fe15a5e-9ee6-47ba-a450-044b62740488","Type":"ContainerStarted","Data":"7f92d5aaefeada9eb78afef7f6706ddb9c30f3bd598b53d70bfdf5ea6256cd85"} Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.937869 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a313c68a-aeef-4003-a77d-97fb7ba6b22c","Type":"ContainerStarted","Data":"ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1"} Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.937902 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a313c68a-aeef-4003-a77d-97fb7ba6b22c","Type":"ContainerStarted","Data":"36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1"} Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.940023 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"426135f0-8fe9-40e8-9a1d-f95c000de11a","Type":"ContainerStarted","Data":"463df52f01d8665c2d960be5d6a0bafa81b877c3ce9fde41cf2bbd3ab37dd67b"} Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.940240 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.959982 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9599668980000002 podStartE2EDuration="2.959966898s" podCreationTimestamp="2025-10-07 23:45:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:24.955336275 +0000 UTC m=+5798.758034368" watchObservedRunningTime="2025-10-07 23:45:24.959966898 +0000 UTC m=+5798.762664971" Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.981115 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9810978759999998 podStartE2EDuration="2.981097876s" podCreationTimestamp="2025-10-07 23:45:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:24.974418449 +0000 UTC m=+5798.777116542" watchObservedRunningTime="2025-10-07 23:45:24.981097876 +0000 UTC m=+5798.783795949" Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.994036 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f23c5ed1-6597-4d11-9270-6c0428a37d42" path="/var/lib/kubelet/pods/f23c5ed1-6597-4d11-9270-6c0428a37d42/volumes" Oct 07 23:45:24 crc kubenswrapper[4871]: I1007 23:45:24.997822 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.997774576 podStartE2EDuration="2.997774576s" podCreationTimestamp="2025-10-07 23:45:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:24.992176458 +0000 UTC m=+5798.794874531" watchObservedRunningTime="2025-10-07 23:45:24.997774576 +0000 UTC m=+5798.800472649" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.330982 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.424211 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b4b8d7867-dqbll"] Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.424503 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" podUID="20c7f7e8-8bea-4384-a21c-4ed818883882" containerName="dnsmasq-dns" containerID="cri-o://57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a" gracePeriod=10 Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.616557 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.678690 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-config-data\") pod \"5e592305-57bf-441b-926e-0217c88ab964\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.678727 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4brhs\" (UniqueName: \"kubernetes.io/projected/5e592305-57bf-441b-926e-0217c88ab964-kube-api-access-4brhs\") pod \"5e592305-57bf-441b-926e-0217c88ab964\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.678785 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-combined-ca-bundle\") pod \"5e592305-57bf-441b-926e-0217c88ab964\" (UID: \"5e592305-57bf-441b-926e-0217c88ab964\") " Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.693337 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e592305-57bf-441b-926e-0217c88ab964-kube-api-access-4brhs" (OuterVolumeSpecName: "kube-api-access-4brhs") pod "5e592305-57bf-441b-926e-0217c88ab964" (UID: "5e592305-57bf-441b-926e-0217c88ab964"). InnerVolumeSpecName "kube-api-access-4brhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.730929 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e592305-57bf-441b-926e-0217c88ab964" (UID: "5e592305-57bf-441b-926e-0217c88ab964"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.773987 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-config-data" (OuterVolumeSpecName: "config-data") pod "5e592305-57bf-441b-926e-0217c88ab964" (UID: "5e592305-57bf-441b-926e-0217c88ab964"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.781128 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.781168 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4brhs\" (UniqueName: \"kubernetes.io/projected/5e592305-57bf-441b-926e-0217c88ab964-kube-api-access-4brhs\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.781179 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e592305-57bf-441b-926e-0217c88ab964-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.956298 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.958115 4871 generic.go:334] "Generic (PLEG): container finished" podID="20c7f7e8-8bea-4384-a21c-4ed818883882" containerID="57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a" exitCode=0 Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.958196 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" event={"ID":"20c7f7e8-8bea-4384-a21c-4ed818883882","Type":"ContainerDied","Data":"57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a"} Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.958453 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" event={"ID":"20c7f7e8-8bea-4384-a21c-4ed818883882","Type":"ContainerDied","Data":"c58ec01186372f44c2e24edf2b6591ad454e2a66715b4fde0358a3fbd4df6777"} Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.958502 4871 scope.go:117] "RemoveContainer" containerID="57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.960038 4871 generic.go:334] "Generic (PLEG): container finished" podID="5e592305-57bf-441b-926e-0217c88ab964" containerID="ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b" exitCode=0 Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.960229 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5e592305-57bf-441b-926e-0217c88ab964","Type":"ContainerDied","Data":"ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b"} Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.960354 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.979188 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5e592305-57bf-441b-926e-0217c88ab964","Type":"ContainerDied","Data":"b0cd581e551a5710791dbe47f87497cbb922b52f595ebd96cabe01be0c8f37fb"} Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.986975 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-sb\") pod \"20c7f7e8-8bea-4384-a21c-4ed818883882\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.987031 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-dns-svc\") pod \"20c7f7e8-8bea-4384-a21c-4ed818883882\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.987081 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-nb\") pod \"20c7f7e8-8bea-4384-a21c-4ed818883882\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.987154 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-config\") pod \"20c7f7e8-8bea-4384-a21c-4ed818883882\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.987239 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq7gm\" (UniqueName: \"kubernetes.io/projected/20c7f7e8-8bea-4384-a21c-4ed818883882-kube-api-access-mq7gm\") pod \"20c7f7e8-8bea-4384-a21c-4ed818883882\" (UID: \"20c7f7e8-8bea-4384-a21c-4ed818883882\") " Oct 07 23:45:25 crc kubenswrapper[4871]: I1007 23:45:25.987973 4871 scope.go:117] "RemoveContainer" containerID="b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.002436 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20c7f7e8-8bea-4384-a21c-4ed818883882-kube-api-access-mq7gm" (OuterVolumeSpecName: "kube-api-access-mq7gm") pod "20c7f7e8-8bea-4384-a21c-4ed818883882" (UID: "20c7f7e8-8bea-4384-a21c-4ed818883882"). InnerVolumeSpecName "kube-api-access-mq7gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.049830 4871 scope.go:117] "RemoveContainer" containerID="57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a" Oct 07 23:45:26 crc kubenswrapper[4871]: E1007 23:45:26.050423 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a\": container with ID starting with 57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a not found: ID does not exist" containerID="57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.050471 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a"} err="failed to get container status \"57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a\": rpc error: code = NotFound desc = could not find container \"57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a\": container with ID starting with 57c36e405dfff944b0cd1e23a9ab85ca0240cd51d217e1db05960967c255635a not found: ID does not exist" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.050514 4871 scope.go:117] "RemoveContainer" containerID="b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619" Oct 07 23:45:26 crc kubenswrapper[4871]: E1007 23:45:26.051600 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619\": container with ID starting with b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619 not found: ID does not exist" containerID="b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.051670 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619"} err="failed to get container status \"b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619\": rpc error: code = NotFound desc = could not find container \"b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619\": container with ID starting with b4fc6bc70cd29e0fe41c217cac5128def3badca84021f73f362fa2bed200a619 not found: ID does not exist" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.051708 4871 scope.go:117] "RemoveContainer" containerID="ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.063526 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "20c7f7e8-8bea-4384-a21c-4ed818883882" (UID: "20c7f7e8-8bea-4384-a21c-4ed818883882"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.065572 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-config" (OuterVolumeSpecName: "config") pod "20c7f7e8-8bea-4384-a21c-4ed818883882" (UID: "20c7f7e8-8bea-4384-a21c-4ed818883882"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.079164 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.086841 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.096423 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "20c7f7e8-8bea-4384-a21c-4ed818883882" (UID: "20c7f7e8-8bea-4384-a21c-4ed818883882"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.096492 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:45:26 crc kubenswrapper[4871]: E1007 23:45:26.096912 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e592305-57bf-441b-926e-0217c88ab964" containerName="nova-scheduler-scheduler" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.096930 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e592305-57bf-441b-926e-0217c88ab964" containerName="nova-scheduler-scheduler" Oct 07 23:45:26 crc kubenswrapper[4871]: E1007 23:45:26.096949 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c7f7e8-8bea-4384-a21c-4ed818883882" containerName="dnsmasq-dns" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.096957 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c7f7e8-8bea-4384-a21c-4ed818883882" containerName="dnsmasq-dns" Oct 07 23:45:26 crc kubenswrapper[4871]: E1007 23:45:26.096978 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c7f7e8-8bea-4384-a21c-4ed818883882" containerName="init" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.096986 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c7f7e8-8bea-4384-a21c-4ed818883882" containerName="init" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.097154 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e592305-57bf-441b-926e-0217c88ab964" containerName="nova-scheduler-scheduler" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.097180 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="20c7f7e8-8bea-4384-a21c-4ed818883882" containerName="dnsmasq-dns" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.097786 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.100704 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.100740 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq7gm\" (UniqueName: \"kubernetes.io/projected/20c7f7e8-8bea-4384-a21c-4ed818883882-kube-api-access-mq7gm\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.100752 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.100760 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.102396 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.103677 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.128110 4871 scope.go:117] "RemoveContainer" containerID="ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b" Oct 07 23:45:26 crc kubenswrapper[4871]: E1007 23:45:26.128886 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b\": container with ID starting with ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b not found: ID does not exist" containerID="ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.128923 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b"} err="failed to get container status \"ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b\": rpc error: code = NotFound desc = could not find container \"ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b\": container with ID starting with ef79524cde27fafd0224a7864d00abb7d62cdf00e582ae2c729825f5a729713b not found: ID does not exist" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.136515 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "20c7f7e8-8bea-4384-a21c-4ed818883882" (UID: "20c7f7e8-8bea-4384-a21c-4ed818883882"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.199403 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.202261 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-config-data\") pod \"nova-scheduler-0\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.202323 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqmqn\" (UniqueName: \"kubernetes.io/projected/a360e652-467f-4d96-9232-d944191ee751-kube-api-access-fqmqn\") pod \"nova-scheduler-0\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.202367 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.202413 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c7f7e8-8bea-4384-a21c-4ed818883882-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.303223 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqmqn\" (UniqueName: \"kubernetes.io/projected/a360e652-467f-4d96-9232-d944191ee751-kube-api-access-fqmqn\") pod \"nova-scheduler-0\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.303292 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.303392 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-config-data\") pod \"nova-scheduler-0\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.307155 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-config-data\") pod \"nova-scheduler-0\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.307653 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.323663 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqmqn\" (UniqueName: \"kubernetes.io/projected/a360e652-467f-4d96-9232-d944191ee751-kube-api-access-fqmqn\") pod \"nova-scheduler-0\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.340591 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.404789 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-combined-ca-bundle\") pod \"88b2fbdc-ee2b-4803-be96-c47713900f08\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.404970 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdffk\" (UniqueName: \"kubernetes.io/projected/88b2fbdc-ee2b-4803-be96-c47713900f08-kube-api-access-tdffk\") pod \"88b2fbdc-ee2b-4803-be96-c47713900f08\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.405048 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-config-data\") pod \"88b2fbdc-ee2b-4803-be96-c47713900f08\" (UID: \"88b2fbdc-ee2b-4803-be96-c47713900f08\") " Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.408271 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b2fbdc-ee2b-4803-be96-c47713900f08-kube-api-access-tdffk" (OuterVolumeSpecName: "kube-api-access-tdffk") pod "88b2fbdc-ee2b-4803-be96-c47713900f08" (UID: "88b2fbdc-ee2b-4803-be96-c47713900f08"). InnerVolumeSpecName "kube-api-access-tdffk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.421183 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.429187 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88b2fbdc-ee2b-4803-be96-c47713900f08" (UID: "88b2fbdc-ee2b-4803-be96-c47713900f08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.432363 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-config-data" (OuterVolumeSpecName: "config-data") pod "88b2fbdc-ee2b-4803-be96-c47713900f08" (UID: "88b2fbdc-ee2b-4803-be96-c47713900f08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.507428 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdffk\" (UniqueName: \"kubernetes.io/projected/88b2fbdc-ee2b-4803-be96-c47713900f08-kube-api-access-tdffk\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.507463 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.507473 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b2fbdc-ee2b-4803-be96-c47713900f08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:26 crc kubenswrapper[4871]: I1007 23:45:26.900290 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.002664 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e592305-57bf-441b-926e-0217c88ab964" path="/var/lib/kubelet/pods/5e592305-57bf-441b-926e-0217c88ab964/volumes" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.004243 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a360e652-467f-4d96-9232-d944191ee751","Type":"ContainerStarted","Data":"47e2e8e455a3610896404f480ee6449179b946796104b0129f381165c5dc47e5"} Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.010436 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.010409 4871 generic.go:334] "Generic (PLEG): container finished" podID="88b2fbdc-ee2b-4803-be96-c47713900f08" containerID="b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a" exitCode=0 Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.010517 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"88b2fbdc-ee2b-4803-be96-c47713900f08","Type":"ContainerDied","Data":"b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a"} Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.010576 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"88b2fbdc-ee2b-4803-be96-c47713900f08","Type":"ContainerDied","Data":"d7b7d7bfe6b672bc79a7bd2efa33caa19d2ae7374b942d25b938feda4ad01e6e"} Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.010599 4871 scope.go:117] "RemoveContainer" containerID="b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.012243 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4b8d7867-dqbll" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.069118 4871 scope.go:117] "RemoveContainer" containerID="b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a" Oct 07 23:45:27 crc kubenswrapper[4871]: E1007 23:45:27.069634 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a\": container with ID starting with b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a not found: ID does not exist" containerID="b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.069671 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a"} err="failed to get container status \"b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a\": rpc error: code = NotFound desc = could not find container \"b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a\": container with ID starting with b8466eb4e388f31a1cde18d84a1cfe0366f6ef8ee7f7722fbc2a927762f8648a not found: ID does not exist" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.099645 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.115634 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.125783 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b4b8d7867-dqbll"] Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.135651 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 23:45:27 crc kubenswrapper[4871]: E1007 23:45:27.136095 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b2fbdc-ee2b-4803-be96-c47713900f08" containerName="nova-cell1-conductor-conductor" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.136113 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b2fbdc-ee2b-4803-be96-c47713900f08" containerName="nova-cell1-conductor-conductor" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.148430 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b2fbdc-ee2b-4803-be96-c47713900f08" containerName="nova-cell1-conductor-conductor" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.149087 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b4b8d7867-dqbll"] Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.149183 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.151425 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.152595 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.220693 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n72pd\" (UniqueName: \"kubernetes.io/projected/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-kube-api-access-n72pd\") pod \"nova-cell1-conductor-0\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.221319 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.221405 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.324706 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.324971 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.325191 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n72pd\" (UniqueName: \"kubernetes.io/projected/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-kube-api-access-n72pd\") pod \"nova-cell1-conductor-0\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.330556 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.331273 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.339631 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n72pd\" (UniqueName: \"kubernetes.io/projected/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-kube-api-access-n72pd\") pod \"nova-cell1-conductor-0\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.346603 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.465259 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:27 crc kubenswrapper[4871]: I1007 23:45:27.919071 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 23:45:28 crc kubenswrapper[4871]: I1007 23:45:28.035695 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a360e652-467f-4d96-9232-d944191ee751","Type":"ContainerStarted","Data":"dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34"} Oct 07 23:45:28 crc kubenswrapper[4871]: I1007 23:45:28.037160 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6e94c8d0-f89e-4215-8c9c-158cfead9a4a","Type":"ContainerStarted","Data":"fc1d17ff6504d420075fc6b882eb1540d2c6cb6d01e78ed391f9c6dad673d2e8"} Oct 07 23:45:28 crc kubenswrapper[4871]: I1007 23:45:28.063225 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.063205751 podStartE2EDuration="2.063205751s" podCreationTimestamp="2025-10-07 23:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:28.05750686 +0000 UTC m=+5801.860204933" watchObservedRunningTime="2025-10-07 23:45:28.063205751 +0000 UTC m=+5801.865903844" Oct 07 23:45:28 crc kubenswrapper[4871]: I1007 23:45:28.358242 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 23:45:28 crc kubenswrapper[4871]: I1007 23:45:28.358459 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 23:45:28 crc kubenswrapper[4871]: I1007 23:45:28.996508 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20c7f7e8-8bea-4384-a21c-4ed818883882" path="/var/lib/kubelet/pods/20c7f7e8-8bea-4384-a21c-4ed818883882/volumes" Oct 07 23:45:28 crc kubenswrapper[4871]: I1007 23:45:28.997342 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88b2fbdc-ee2b-4803-be96-c47713900f08" path="/var/lib/kubelet/pods/88b2fbdc-ee2b-4803-be96-c47713900f08/volumes" Oct 07 23:45:29 crc kubenswrapper[4871]: I1007 23:45:29.051209 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6e94c8d0-f89e-4215-8c9c-158cfead9a4a","Type":"ContainerStarted","Data":"6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221"} Oct 07 23:45:29 crc kubenswrapper[4871]: I1007 23:45:29.051270 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:29 crc kubenswrapper[4871]: I1007 23:45:29.073978 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.073955053 podStartE2EDuration="2.073955053s" podCreationTimestamp="2025-10-07 23:45:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:29.067541304 +0000 UTC m=+5802.870239387" watchObservedRunningTime="2025-10-07 23:45:29.073955053 +0000 UTC m=+5802.876653136" Oct 07 23:45:31 crc kubenswrapper[4871]: I1007 23:45:31.194364 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:31 crc kubenswrapper[4871]: I1007 23:45:31.209990 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:31 crc kubenswrapper[4871]: I1007 23:45:31.421712 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 23:45:32 crc kubenswrapper[4871]: I1007 23:45:32.095673 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 07 23:45:33 crc kubenswrapper[4871]: I1007 23:45:33.298991 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 23:45:33 crc kubenswrapper[4871]: I1007 23:45:33.299491 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 23:45:33 crc kubenswrapper[4871]: I1007 23:45:33.359041 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 23:45:33 crc kubenswrapper[4871]: I1007 23:45:33.359091 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 23:45:33 crc kubenswrapper[4871]: I1007 23:45:33.360321 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 07 23:45:34 crc kubenswrapper[4871]: I1007 23:45:34.382992 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.79:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:45:34 crc kubenswrapper[4871]: I1007 23:45:34.383041 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.79:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:45:34 crc kubenswrapper[4871]: I1007 23:45:34.465979 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:45:34 crc kubenswrapper[4871]: I1007 23:45:34.466043 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 23:45:34 crc kubenswrapper[4871]: I1007 23:45:34.982533 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:45:34 crc kubenswrapper[4871]: E1007 23:45:34.982978 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:45:36 crc kubenswrapper[4871]: I1007 23:45:36.422432 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 23:45:36 crc kubenswrapper[4871]: I1007 23:45:36.458483 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 23:45:37 crc kubenswrapper[4871]: I1007 23:45:37.180460 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 23:45:37 crc kubenswrapper[4871]: I1007 23:45:37.510154 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 07 23:45:43 crc kubenswrapper[4871]: I1007 23:45:43.305784 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 23:45:43 crc kubenswrapper[4871]: I1007 23:45:43.307121 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 23:45:43 crc kubenswrapper[4871]: I1007 23:45:43.309005 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 23:45:43 crc kubenswrapper[4871]: I1007 23:45:43.324074 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 23:45:43 crc kubenswrapper[4871]: I1007 23:45:43.386328 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 23:45:43 crc kubenswrapper[4871]: I1007 23:45:43.386605 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 23:45:43 crc kubenswrapper[4871]: I1007 23:45:43.396186 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 23:45:43 crc kubenswrapper[4871]: I1007 23:45:43.396271 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 23:45:44 crc kubenswrapper[4871]: I1007 23:45:44.233397 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 23:45:44 crc kubenswrapper[4871]: I1007 23:45:44.239288 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.137681 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.140146 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.143346 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.162400 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.253557 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-scripts\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.253645 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.253730 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.253885 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bghm6\" (UniqueName: \"kubernetes.io/projected/2d1d4a3f-48b4-4466-9943-46832656da24-kube-api-access-bghm6\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.253966 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1d4a3f-48b4-4466-9943-46832656da24-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.254050 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.356115 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-scripts\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.356625 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.356654 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.356685 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bghm6\" (UniqueName: \"kubernetes.io/projected/2d1d4a3f-48b4-4466-9943-46832656da24-kube-api-access-bghm6\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.356710 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1d4a3f-48b4-4466-9943-46832656da24-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.356754 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.357310 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1d4a3f-48b4-4466-9943-46832656da24-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.362725 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-scripts\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.363041 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.363363 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.365164 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.385261 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bghm6\" (UniqueName: \"kubernetes.io/projected/2d1d4a3f-48b4-4466-9943-46832656da24-kube-api-access-bghm6\") pod \"cinder-scheduler-0\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.490219 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 23:45:46 crc kubenswrapper[4871]: I1007 23:45:46.948882 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 23:45:47 crc kubenswrapper[4871]: I1007 23:45:47.266730 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2d1d4a3f-48b4-4466-9943-46832656da24","Type":"ContainerStarted","Data":"bc4ae74c68e00901168250b5c70a52587123cf9e68da3674b24df915055ed899"} Oct 07 23:45:47 crc kubenswrapper[4871]: I1007 23:45:47.323230 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 23:45:47 crc kubenswrapper[4871]: I1007 23:45:47.323456 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerName="cinder-api-log" containerID="cri-o://80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67" gracePeriod=30 Oct 07 23:45:47 crc kubenswrapper[4871]: I1007 23:45:47.323578 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerName="cinder-api" containerID="cri-o://d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c" gracePeriod=30 Oct 07 23:45:47 crc kubenswrapper[4871]: I1007 23:45:47.916699 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 07 23:45:47 crc kubenswrapper[4871]: I1007 23:45:47.923629 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:47 crc kubenswrapper[4871]: I1007 23:45:47.928126 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 07 23:45:47 crc kubenswrapper[4871]: I1007 23:45:47.996873 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002532 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002617 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002641 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002665 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002692 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-run\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002732 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002757 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002777 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002803 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002816 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-sys\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002844 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-dev\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002865 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002890 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60ebc4f7-1bb0-481e-871e-a36c56c744a2-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002907 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002926 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.002956 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b86g\" (UniqueName: \"kubernetes.io/projected/60ebc4f7-1bb0-481e-871e-a36c56c744a2-kube-api-access-4b86g\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106061 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106128 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106144 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-sys\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106168 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-dev\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106196 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106242 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60ebc4f7-1bb0-481e-871e-a36c56c744a2-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106256 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106272 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106309 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b86g\" (UniqueName: \"kubernetes.io/projected/60ebc4f7-1bb0-481e-871e-a36c56c744a2-kube-api-access-4b86g\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106327 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106396 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106423 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106443 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106464 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106533 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-run\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106485 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-run\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106701 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106780 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106940 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.106940 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-sys\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.107021 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.107413 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.107666 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.107688 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.107903 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-dev\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.107918 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/60ebc4f7-1bb0-481e-871e-a36c56c744a2-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.111366 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.111473 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.111861 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.122786 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60ebc4f7-1bb0-481e-871e-a36c56c744a2-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.123237 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ebc4f7-1bb0-481e-871e-a36c56c744a2-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.125214 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b86g\" (UniqueName: \"kubernetes.io/projected/60ebc4f7-1bb0-481e-871e-a36c56c744a2-kube-api-access-4b86g\") pod \"cinder-volume-volume1-0\" (UID: \"60ebc4f7-1bb0-481e-871e-a36c56c744a2\") " pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.277674 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2d1d4a3f-48b4-4466-9943-46832656da24","Type":"ContainerStarted","Data":"7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662"} Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.279906 4871 generic.go:334] "Generic (PLEG): container finished" podID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerID="80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67" exitCode=143 Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.279936 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af866d9d-14f4-4e78-82f9-01102c64aea5","Type":"ContainerDied","Data":"80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67"} Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.321571 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.568557 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.570498 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.578990 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.586332 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.721340 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.721389 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-lib-modules\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.721495 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/06fc942e-87b3-435b-97c3-4a42d7d78abe-ceph\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.721787 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-dev\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.721892 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-config-data-custom\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.721929 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-etc-nvme\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.721950 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.721965 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-config-data\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.722015 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-run\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.722033 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-sys\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.722060 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.722078 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.722095 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r882j\" (UniqueName: \"kubernetes.io/projected/06fc942e-87b3-435b-97c3-4a42d7d78abe-kube-api-access-r882j\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.722116 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.722140 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.722156 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-scripts\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.823636 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-dev\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.823704 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-config-data-custom\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.823742 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-etc-nvme\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.823776 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.823804 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-config-data\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.823787 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-dev\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.823975 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-run\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824015 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-sys\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824086 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824131 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r882j\" (UniqueName: \"kubernetes.io/projected/06fc942e-87b3-435b-97c3-4a42d7d78abe-kube-api-access-r882j\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824158 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824194 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824237 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824263 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-scripts\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824384 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824412 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-lib-modules\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824455 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/06fc942e-87b3-435b-97c3-4a42d7d78abe-ceph\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824585 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824721 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-lib-modules\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824765 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-etc-nvme\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824759 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.824799 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.825160 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.825202 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.825235 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-sys\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.825258 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/06fc942e-87b3-435b-97c3-4a42d7d78abe-run\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.828734 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-scripts\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.829044 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.830144 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-config-data\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.843399 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/06fc942e-87b3-435b-97c3-4a42d7d78abe-ceph\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.843841 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06fc942e-87b3-435b-97c3-4a42d7d78abe-config-data-custom\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.851721 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r882j\" (UniqueName: \"kubernetes.io/projected/06fc942e-87b3-435b-97c3-4a42d7d78abe-kube-api-access-r882j\") pod \"cinder-backup-0\" (UID: \"06fc942e-87b3-435b-97c3-4a42d7d78abe\") " pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.889094 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.912474 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 07 23:45:48 crc kubenswrapper[4871]: I1007 23:45:48.982859 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:45:48 crc kubenswrapper[4871]: E1007 23:45:48.983383 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:45:49 crc kubenswrapper[4871]: I1007 23:45:49.290970 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2d1d4a3f-48b4-4466-9943-46832656da24","Type":"ContainerStarted","Data":"c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f"} Oct 07 23:45:49 crc kubenswrapper[4871]: I1007 23:45:49.293059 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"60ebc4f7-1bb0-481e-871e-a36c56c744a2","Type":"ContainerStarted","Data":"45f190022ad034b5a6e3f244974c7c5d4de8b8025f538286d7685a2c138fdfaa"} Oct 07 23:45:49 crc kubenswrapper[4871]: I1007 23:45:49.312321 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.31229166 podStartE2EDuration="3.31229166s" podCreationTimestamp="2025-10-07 23:45:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:49.309500666 +0000 UTC m=+5823.112198739" watchObservedRunningTime="2025-10-07 23:45:49.31229166 +0000 UTC m=+5823.114989773" Oct 07 23:45:49 crc kubenswrapper[4871]: I1007 23:45:49.440250 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.303889 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"06fc942e-87b3-435b-97c3-4a42d7d78abe","Type":"ContainerStarted","Data":"dc9eaa0e7fea2fa5dc3c1bb24386f85683438b9c7bbcf848f5b111c7ffd76fda"} Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.305514 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"60ebc4f7-1bb0-481e-871e-a36c56c744a2","Type":"ContainerStarted","Data":"426d1c133f31efeb65bb5e578959a84bb10ece1f1f3aeff0819f507860177540"} Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.464228 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.77:8776/healthcheck\": read tcp 10.217.0.2:44332->10.217.1.77:8776: read: connection reset by peer" Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.866005 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.973236 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-scripts\") pod \"af866d9d-14f4-4e78-82f9-01102c64aea5\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.973592 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af866d9d-14f4-4e78-82f9-01102c64aea5-etc-machine-id\") pod \"af866d9d-14f4-4e78-82f9-01102c64aea5\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.973630 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af866d9d-14f4-4e78-82f9-01102c64aea5-logs\") pod \"af866d9d-14f4-4e78-82f9-01102c64aea5\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.973677 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-combined-ca-bundle\") pod \"af866d9d-14f4-4e78-82f9-01102c64aea5\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.973731 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpf65\" (UniqueName: \"kubernetes.io/projected/af866d9d-14f4-4e78-82f9-01102c64aea5-kube-api-access-wpf65\") pod \"af866d9d-14f4-4e78-82f9-01102c64aea5\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.973912 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data-custom\") pod \"af866d9d-14f4-4e78-82f9-01102c64aea5\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.973947 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data\") pod \"af866d9d-14f4-4e78-82f9-01102c64aea5\" (UID: \"af866d9d-14f4-4e78-82f9-01102c64aea5\") " Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.975223 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af866d9d-14f4-4e78-82f9-01102c64aea5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "af866d9d-14f4-4e78-82f9-01102c64aea5" (UID: "af866d9d-14f4-4e78-82f9-01102c64aea5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.975603 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af866d9d-14f4-4e78-82f9-01102c64aea5-logs" (OuterVolumeSpecName: "logs") pod "af866d9d-14f4-4e78-82f9-01102c64aea5" (UID: "af866d9d-14f4-4e78-82f9-01102c64aea5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.979625 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "af866d9d-14f4-4e78-82f9-01102c64aea5" (UID: "af866d9d-14f4-4e78-82f9-01102c64aea5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.979649 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af866d9d-14f4-4e78-82f9-01102c64aea5-kube-api-access-wpf65" (OuterVolumeSpecName: "kube-api-access-wpf65") pod "af866d9d-14f4-4e78-82f9-01102c64aea5" (UID: "af866d9d-14f4-4e78-82f9-01102c64aea5"). InnerVolumeSpecName "kube-api-access-wpf65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:50 crc kubenswrapper[4871]: I1007 23:45:50.980370 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-scripts" (OuterVolumeSpecName: "scripts") pod "af866d9d-14f4-4e78-82f9-01102c64aea5" (UID: "af866d9d-14f4-4e78-82f9-01102c64aea5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.006697 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af866d9d-14f4-4e78-82f9-01102c64aea5" (UID: "af866d9d-14f4-4e78-82f9-01102c64aea5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.023867 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data" (OuterVolumeSpecName: "config-data") pod "af866d9d-14f4-4e78-82f9-01102c64aea5" (UID: "af866d9d-14f4-4e78-82f9-01102c64aea5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.076280 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.076314 4871 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af866d9d-14f4-4e78-82f9-01102c64aea5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.076324 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af866d9d-14f4-4e78-82f9-01102c64aea5-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.076336 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.076345 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpf65\" (UniqueName: \"kubernetes.io/projected/af866d9d-14f4-4e78-82f9-01102c64aea5-kube-api-access-wpf65\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.076357 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.076365 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af866d9d-14f4-4e78-82f9-01102c64aea5-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.320492 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"06fc942e-87b3-435b-97c3-4a42d7d78abe","Type":"ContainerStarted","Data":"b5f61a72567e78d2137be8959e0ada88a57f2e2cdbaad3fdb088be67a3fa4d0f"} Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.320533 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"06fc942e-87b3-435b-97c3-4a42d7d78abe","Type":"ContainerStarted","Data":"92fc72a3b584ef07a59a898f3fabe1d3b0d8985e9ec1aea997925db45507d66d"} Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.326479 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"60ebc4f7-1bb0-481e-871e-a36c56c744a2","Type":"ContainerStarted","Data":"4a46f4c32eff2ac0bd75c5658210ea2dd10f64122744abe8f059fba132ffaebe"} Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.329635 4871 generic.go:334] "Generic (PLEG): container finished" podID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerID="d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c" exitCode=0 Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.329670 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af866d9d-14f4-4e78-82f9-01102c64aea5","Type":"ContainerDied","Data":"d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c"} Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.329693 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"af866d9d-14f4-4e78-82f9-01102c64aea5","Type":"ContainerDied","Data":"520a6f412c4bba09a2941757238cd87336cf860ec61ec04efe5c630909b7c189"} Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.329707 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.329729 4871 scope.go:117] "RemoveContainer" containerID="d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.348426 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.477627553 podStartE2EDuration="3.348407351s" podCreationTimestamp="2025-10-07 23:45:48 +0000 UTC" firstStartedPulling="2025-10-07 23:45:49.438483511 +0000 UTC m=+5823.241181574" lastFinishedPulling="2025-10-07 23:45:50.309263299 +0000 UTC m=+5824.111961372" observedRunningTime="2025-10-07 23:45:51.343329277 +0000 UTC m=+5825.146027340" watchObservedRunningTime="2025-10-07 23:45:51.348407351 +0000 UTC m=+5825.151105454" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.361509 4871 scope.go:117] "RemoveContainer" containerID="80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.385392 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.335283305 podStartE2EDuration="4.385371997s" podCreationTimestamp="2025-10-07 23:45:47 +0000 UTC" firstStartedPulling="2025-10-07 23:45:48.934348352 +0000 UTC m=+5822.737046425" lastFinishedPulling="2025-10-07 23:45:49.984437014 +0000 UTC m=+5823.787135117" observedRunningTime="2025-10-07 23:45:51.384058852 +0000 UTC m=+5825.186756935" watchObservedRunningTime="2025-10-07 23:45:51.385371997 +0000 UTC m=+5825.188070090" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.401406 4871 scope.go:117] "RemoveContainer" containerID="d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c" Oct 07 23:45:51 crc kubenswrapper[4871]: E1007 23:45:51.401749 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c\": container with ID starting with d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c not found: ID does not exist" containerID="d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.401778 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c"} err="failed to get container status \"d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c\": rpc error: code = NotFound desc = could not find container \"d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c\": container with ID starting with d098dcdcbd48e07d364634d6cd087b67a6efa4c93494252f98f496a08b4c000c not found: ID does not exist" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.401800 4871 scope.go:117] "RemoveContainer" containerID="80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67" Oct 07 23:45:51 crc kubenswrapper[4871]: E1007 23:45:51.402383 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67\": container with ID starting with 80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67 not found: ID does not exist" containerID="80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.402436 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67"} err="failed to get container status \"80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67\": rpc error: code = NotFound desc = could not find container \"80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67\": container with ID starting with 80a3b3e52c514c6e008ae5a6db18b6578e3ec8269d885ab0a51c528d74b19f67 not found: ID does not exist" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.417191 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.447889 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.447942 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 07 23:45:51 crc kubenswrapper[4871]: E1007 23:45:51.448274 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerName="cinder-api-log" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.448287 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerName="cinder-api-log" Oct 07 23:45:51 crc kubenswrapper[4871]: E1007 23:45:51.448309 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerName="cinder-api" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.448316 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerName="cinder-api" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.448480 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerName="cinder-api" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.448498 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" containerName="cinder-api-log" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.449321 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.449397 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.466337 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.492311 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.593097 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.593187 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf095ee8-6cee-4ab7-888d-8cea407f2e82-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.593232 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rvfs\" (UniqueName: \"kubernetes.io/projected/bf095ee8-6cee-4ab7-888d-8cea407f2e82-kube-api-access-4rvfs\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.593384 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-config-data-custom\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.593436 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-scripts\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.593565 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-config-data\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.593621 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf095ee8-6cee-4ab7-888d-8cea407f2e82-logs\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.695438 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-config-data\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.695746 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf095ee8-6cee-4ab7-888d-8cea407f2e82-logs\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.695993 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.696133 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf095ee8-6cee-4ab7-888d-8cea407f2e82-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.696248 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rvfs\" (UniqueName: \"kubernetes.io/projected/bf095ee8-6cee-4ab7-888d-8cea407f2e82-kube-api-access-4rvfs\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.696364 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-config-data-custom\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.696479 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-scripts\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.696245 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf095ee8-6cee-4ab7-888d-8cea407f2e82-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.698458 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf095ee8-6cee-4ab7-888d-8cea407f2e82-logs\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.699664 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-config-data\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.699922 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-config-data-custom\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.707153 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-scripts\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.713972 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf095ee8-6cee-4ab7-888d-8cea407f2e82-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.720556 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rvfs\" (UniqueName: \"kubernetes.io/projected/bf095ee8-6cee-4ab7-888d-8cea407f2e82-kube-api-access-4rvfs\") pod \"cinder-api-0\" (UID: \"bf095ee8-6cee-4ab7-888d-8cea407f2e82\") " pod="openstack/cinder-api-0" Oct 07 23:45:51 crc kubenswrapper[4871]: I1007 23:45:51.817870 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 23:45:52 crc kubenswrapper[4871]: I1007 23:45:52.072403 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 23:45:52 crc kubenswrapper[4871]: I1007 23:45:52.344487 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bf095ee8-6cee-4ab7-888d-8cea407f2e82","Type":"ContainerStarted","Data":"8d3012ed558ea9c7e2945378f77a7ba5b895eb087f178b56da23fd4bef7a51d9"} Oct 07 23:45:52 crc kubenswrapper[4871]: I1007 23:45:52.996001 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af866d9d-14f4-4e78-82f9-01102c64aea5" path="/var/lib/kubelet/pods/af866d9d-14f4-4e78-82f9-01102c64aea5/volumes" Oct 07 23:45:53 crc kubenswrapper[4871]: I1007 23:45:53.321753 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:53 crc kubenswrapper[4871]: I1007 23:45:53.371768 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bf095ee8-6cee-4ab7-888d-8cea407f2e82","Type":"ContainerStarted","Data":"0718a98a1741d1906d7f8deea7444cd803c2d1d629bd73f3e8a336939f164d50"} Oct 07 23:45:53 crc kubenswrapper[4871]: I1007 23:45:53.890878 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 07 23:45:54 crc kubenswrapper[4871]: I1007 23:45:54.386582 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bf095ee8-6cee-4ab7-888d-8cea407f2e82","Type":"ContainerStarted","Data":"1e04741e20f93e310115d81ac5014998bac20dc589e38080a0b498fb27139857"} Oct 07 23:45:54 crc kubenswrapper[4871]: I1007 23:45:54.387082 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 07 23:45:54 crc kubenswrapper[4871]: I1007 23:45:54.414942 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.414917064 podStartE2EDuration="3.414917064s" podCreationTimestamp="2025-10-07 23:45:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:45:54.410342134 +0000 UTC m=+5828.213040197" watchObservedRunningTime="2025-10-07 23:45:54.414917064 +0000 UTC m=+5828.217615167" Oct 07 23:45:56 crc kubenswrapper[4871]: I1007 23:45:56.723291 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 07 23:45:56 crc kubenswrapper[4871]: I1007 23:45:56.825738 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 23:45:57 crc kubenswrapper[4871]: I1007 23:45:57.419478 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2d1d4a3f-48b4-4466-9943-46832656da24" containerName="cinder-scheduler" containerID="cri-o://7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662" gracePeriod=30 Oct 07 23:45:57 crc kubenswrapper[4871]: I1007 23:45:57.419604 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2d1d4a3f-48b4-4466-9943-46832656da24" containerName="probe" containerID="cri-o://c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f" gracePeriod=30 Oct 07 23:45:57 crc kubenswrapper[4871]: I1007 23:45:57.608543 4871 scope.go:117] "RemoveContainer" containerID="3f8117e13ab5e1f0122f7cdaa0e7186308a0aa94c58dc8efbcc0f2746b40cc30" Oct 07 23:45:58 crc kubenswrapper[4871]: I1007 23:45:58.433956 4871 generic.go:334] "Generic (PLEG): container finished" podID="2d1d4a3f-48b4-4466-9943-46832656da24" containerID="c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f" exitCode=0 Oct 07 23:45:58 crc kubenswrapper[4871]: I1007 23:45:58.434012 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2d1d4a3f-48b4-4466-9943-46832656da24","Type":"ContainerDied","Data":"c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f"} Oct 07 23:45:58 crc kubenswrapper[4871]: I1007 23:45:58.565893 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 07 23:45:58 crc kubenswrapper[4871]: I1007 23:45:58.944291 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.065221 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bghm6\" (UniqueName: \"kubernetes.io/projected/2d1d4a3f-48b4-4466-9943-46832656da24-kube-api-access-bghm6\") pod \"2d1d4a3f-48b4-4466-9943-46832656da24\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.065288 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1d4a3f-48b4-4466-9943-46832656da24-etc-machine-id\") pod \"2d1d4a3f-48b4-4466-9943-46832656da24\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.065371 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data\") pod \"2d1d4a3f-48b4-4466-9943-46832656da24\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.065444 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data-custom\") pod \"2d1d4a3f-48b4-4466-9943-46832656da24\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.065511 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d1d4a3f-48b4-4466-9943-46832656da24-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2d1d4a3f-48b4-4466-9943-46832656da24" (UID: "2d1d4a3f-48b4-4466-9943-46832656da24"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.065608 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-combined-ca-bundle\") pod \"2d1d4a3f-48b4-4466-9943-46832656da24\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.065673 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-scripts\") pod \"2d1d4a3f-48b4-4466-9943-46832656da24\" (UID: \"2d1d4a3f-48b4-4466-9943-46832656da24\") " Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.066267 4871 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1d4a3f-48b4-4466-9943-46832656da24-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.071054 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d1d4a3f-48b4-4466-9943-46832656da24-kube-api-access-bghm6" (OuterVolumeSpecName: "kube-api-access-bghm6") pod "2d1d4a3f-48b4-4466-9943-46832656da24" (UID: "2d1d4a3f-48b4-4466-9943-46832656da24"). InnerVolumeSpecName "kube-api-access-bghm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.071164 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2d1d4a3f-48b4-4466-9943-46832656da24" (UID: "2d1d4a3f-48b4-4466-9943-46832656da24"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.071237 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-scripts" (OuterVolumeSpecName: "scripts") pod "2d1d4a3f-48b4-4466-9943-46832656da24" (UID: "2d1d4a3f-48b4-4466-9943-46832656da24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.119755 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d1d4a3f-48b4-4466-9943-46832656da24" (UID: "2d1d4a3f-48b4-4466-9943-46832656da24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.124551 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.168451 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.168489 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.168501 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.168513 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bghm6\" (UniqueName: \"kubernetes.io/projected/2d1d4a3f-48b4-4466-9943-46832656da24-kube-api-access-bghm6\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.179783 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data" (OuterVolumeSpecName: "config-data") pod "2d1d4a3f-48b4-4466-9943-46832656da24" (UID: "2d1d4a3f-48b4-4466-9943-46832656da24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.270062 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1d4a3f-48b4-4466-9943-46832656da24-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.445634 4871 generic.go:334] "Generic (PLEG): container finished" podID="2d1d4a3f-48b4-4466-9943-46832656da24" containerID="7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662" exitCode=0 Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.445698 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2d1d4a3f-48b4-4466-9943-46832656da24","Type":"ContainerDied","Data":"7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662"} Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.445780 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2d1d4a3f-48b4-4466-9943-46832656da24","Type":"ContainerDied","Data":"bc4ae74c68e00901168250b5c70a52587123cf9e68da3674b24df915055ed899"} Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.445719 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.445841 4871 scope.go:117] "RemoveContainer" containerID="c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.475890 4871 scope.go:117] "RemoveContainer" containerID="7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.483164 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.508850 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.514771 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 23:45:59 crc kubenswrapper[4871]: E1007 23:45:59.515183 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1d4a3f-48b4-4466-9943-46832656da24" containerName="probe" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.515199 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1d4a3f-48b4-4466-9943-46832656da24" containerName="probe" Oct 07 23:45:59 crc kubenswrapper[4871]: E1007 23:45:59.515230 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1d4a3f-48b4-4466-9943-46832656da24" containerName="cinder-scheduler" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.515236 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1d4a3f-48b4-4466-9943-46832656da24" containerName="cinder-scheduler" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.515448 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d1d4a3f-48b4-4466-9943-46832656da24" containerName="cinder-scheduler" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.515466 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d1d4a3f-48b4-4466-9943-46832656da24" containerName="probe" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.517005 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.520273 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.534006 4871 scope.go:117] "RemoveContainer" containerID="c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f" Oct 07 23:45:59 crc kubenswrapper[4871]: E1007 23:45:59.534442 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f\": container with ID starting with c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f not found: ID does not exist" containerID="c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.534492 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f"} err="failed to get container status \"c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f\": rpc error: code = NotFound desc = could not find container \"c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f\": container with ID starting with c45ccf71e528e2294600fa127c04c9c7eea1a4839d481966ed0bd0f1f461e15f not found: ID does not exist" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.534526 4871 scope.go:117] "RemoveContainer" containerID="7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662" Oct 07 23:45:59 crc kubenswrapper[4871]: E1007 23:45:59.534960 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662\": container with ID starting with 7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662 not found: ID does not exist" containerID="7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.534981 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662"} err="failed to get container status \"7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662\": rpc error: code = NotFound desc = could not find container \"7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662\": container with ID starting with 7196d1dddfc200131a884d5736543cc8db1073b10e87bc2c577a52c9d2908662 not found: ID does not exist" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.550416 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.676352 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.676446 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/213171d1-7fd7-4b97-84cb-7917496a94a4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.676837 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-scripts\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.676959 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqnj8\" (UniqueName: \"kubernetes.io/projected/213171d1-7fd7-4b97-84cb-7917496a94a4-kube-api-access-dqnj8\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.677242 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.677310 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-config-data\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.779203 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.779279 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-config-data\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.779462 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.779527 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/213171d1-7fd7-4b97-84cb-7917496a94a4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.779653 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-scripts\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.779708 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqnj8\" (UniqueName: \"kubernetes.io/projected/213171d1-7fd7-4b97-84cb-7917496a94a4-kube-api-access-dqnj8\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.779731 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/213171d1-7fd7-4b97-84cb-7917496a94a4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.784940 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.786606 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.786927 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-scripts\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.789310 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/213171d1-7fd7-4b97-84cb-7917496a94a4-config-data\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.802006 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqnj8\" (UniqueName: \"kubernetes.io/projected/213171d1-7fd7-4b97-84cb-7917496a94a4-kube-api-access-dqnj8\") pod \"cinder-scheduler-0\" (UID: \"213171d1-7fd7-4b97-84cb-7917496a94a4\") " pod="openstack/cinder-scheduler-0" Oct 07 23:45:59 crc kubenswrapper[4871]: I1007 23:45:59.846086 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 23:46:00 crc kubenswrapper[4871]: I1007 23:46:00.316302 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 23:46:00 crc kubenswrapper[4871]: W1007 23:46:00.325698 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod213171d1_7fd7_4b97_84cb_7917496a94a4.slice/crio-8bad287241b39a9b572b5d7d319f9453a3f2d01ed41e6693119b667e97bcb85f WatchSource:0}: Error finding container 8bad287241b39a9b572b5d7d319f9453a3f2d01ed41e6693119b667e97bcb85f: Status 404 returned error can't find the container with id 8bad287241b39a9b572b5d7d319f9453a3f2d01ed41e6693119b667e97bcb85f Oct 07 23:46:00 crc kubenswrapper[4871]: I1007 23:46:00.458467 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"213171d1-7fd7-4b97-84cb-7917496a94a4","Type":"ContainerStarted","Data":"8bad287241b39a9b572b5d7d319f9453a3f2d01ed41e6693119b667e97bcb85f"} Oct 07 23:46:00 crc kubenswrapper[4871]: I1007 23:46:00.999295 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d1d4a3f-48b4-4466-9943-46832656da24" path="/var/lib/kubelet/pods/2d1d4a3f-48b4-4466-9943-46832656da24/volumes" Oct 07 23:46:01 crc kubenswrapper[4871]: I1007 23:46:01.469693 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"213171d1-7fd7-4b97-84cb-7917496a94a4","Type":"ContainerStarted","Data":"bd5b9d73553d71a043bf80c2c841cefeae40d8bec2da0151e44c60ae1b483f07"} Oct 07 23:46:02 crc kubenswrapper[4871]: I1007 23:46:02.484066 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"213171d1-7fd7-4b97-84cb-7917496a94a4","Type":"ContainerStarted","Data":"2bdfc72579a8902d6d652a4837d7bce07bcb7843b32ce4f3e93d26e564d1dee3"} Oct 07 23:46:02 crc kubenswrapper[4871]: I1007 23:46:02.524403 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.5243762480000003 podStartE2EDuration="3.524376248s" podCreationTimestamp="2025-10-07 23:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:46:02.503557508 +0000 UTC m=+5836.306255581" watchObservedRunningTime="2025-10-07 23:46:02.524376248 +0000 UTC m=+5836.327074361" Oct 07 23:46:02 crc kubenswrapper[4871]: I1007 23:46:02.983286 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:46:02 crc kubenswrapper[4871]: E1007 23:46:02.983760 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:46:03 crc kubenswrapper[4871]: I1007 23:46:03.593834 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 07 23:46:04 crc kubenswrapper[4871]: I1007 23:46:04.846784 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 07 23:46:10 crc kubenswrapper[4871]: I1007 23:46:10.162710 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 07 23:46:15 crc kubenswrapper[4871]: I1007 23:46:15.983020 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:46:15 crc kubenswrapper[4871]: E1007 23:46:15.983890 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:46:28 crc kubenswrapper[4871]: I1007 23:46:28.982740 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:46:28 crc kubenswrapper[4871]: E1007 23:46:28.983930 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:46:31 crc kubenswrapper[4871]: I1007 23:46:31.085061 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-mldkt"] Oct 07 23:46:31 crc kubenswrapper[4871]: I1007 23:46:31.101186 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-mldkt"] Oct 07 23:46:33 crc kubenswrapper[4871]: I1007 23:46:33.019594 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9936e5dd-19e9-4fea-89a2-f4c0f202f880" path="/var/lib/kubelet/pods/9936e5dd-19e9-4fea-89a2-f4c0f202f880/volumes" Oct 07 23:46:41 crc kubenswrapper[4871]: I1007 23:46:41.049017 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-001f-account-create-8vn69"] Oct 07 23:46:41 crc kubenswrapper[4871]: I1007 23:46:41.058260 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-001f-account-create-8vn69"] Oct 07 23:46:41 crc kubenswrapper[4871]: I1007 23:46:41.983539 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:46:41 crc kubenswrapper[4871]: E1007 23:46:41.984653 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:46:43 crc kubenswrapper[4871]: I1007 23:46:43.004837 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6376877b-2033-41cb-832c-9e9740873964" path="/var/lib/kubelet/pods/6376877b-2033-41cb-832c-9e9740873964/volumes" Oct 07 23:46:48 crc kubenswrapper[4871]: I1007 23:46:48.051612 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-f546f"] Oct 07 23:46:48 crc kubenswrapper[4871]: I1007 23:46:48.061537 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-f546f"] Oct 07 23:46:49 crc kubenswrapper[4871]: I1007 23:46:49.001313 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48" path="/var/lib/kubelet/pods/4bbd68b9-4bb4-451a-a35c-48b3c4a6dd48/volumes" Oct 07 23:46:53 crc kubenswrapper[4871]: I1007 23:46:53.982386 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:46:53 crc kubenswrapper[4871]: E1007 23:46:53.983338 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:46:57 crc kubenswrapper[4871]: I1007 23:46:57.826923 4871 scope.go:117] "RemoveContainer" containerID="ace0d887e0e6ec937663f092e2b65b870376917e90f9053593a716352bc8f229" Oct 07 23:46:57 crc kubenswrapper[4871]: I1007 23:46:57.854305 4871 scope.go:117] "RemoveContainer" containerID="c21ce527dcbbdccc9c90446fb3706cecff68b7e4d689683367a00c6cbb1ed7a8" Oct 07 23:46:57 crc kubenswrapper[4871]: I1007 23:46:57.929034 4871 scope.go:117] "RemoveContainer" containerID="90f221112885821cd8b74c5c3e4ff10f80c3c23b867c710289404d1c9753e5e4" Oct 07 23:47:01 crc kubenswrapper[4871]: I1007 23:47:01.050001 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-tm4p2"] Oct 07 23:47:01 crc kubenswrapper[4871]: I1007 23:47:01.059214 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-tm4p2"] Oct 07 23:47:02 crc kubenswrapper[4871]: I1007 23:47:02.998077 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84322a3a-2c29-49e7-a689-4ba61d8916c1" path="/var/lib/kubelet/pods/84322a3a-2c29-49e7-a689-4ba61d8916c1/volumes" Oct 07 23:47:05 crc kubenswrapper[4871]: I1007 23:47:05.983518 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:47:05 crc kubenswrapper[4871]: E1007 23:47:05.983930 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:47:17 crc kubenswrapper[4871]: I1007 23:47:17.982954 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:47:17 crc kubenswrapper[4871]: E1007 23:47:17.985972 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:47:31 crc kubenswrapper[4871]: I1007 23:47:31.982422 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:47:31 crc kubenswrapper[4871]: E1007 23:47:31.983158 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:47:42 crc kubenswrapper[4871]: I1007 23:47:42.983294 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:47:42 crc kubenswrapper[4871]: E1007 23:47:42.984546 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.453290 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jfn58"] Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.457635 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.504095 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jfn58"] Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.566231 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-utilities\") pod \"community-operators-jfn58\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.566660 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-catalog-content\") pod \"community-operators-jfn58\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.566748 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7hjc\" (UniqueName: \"kubernetes.io/projected/30dea74f-070d-464e-8ed5-806cbd07b4f9-kube-api-access-j7hjc\") pod \"community-operators-jfn58\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.668968 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-catalog-content\") pod \"community-operators-jfn58\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.669019 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7hjc\" (UniqueName: \"kubernetes.io/projected/30dea74f-070d-464e-8ed5-806cbd07b4f9-kube-api-access-j7hjc\") pod \"community-operators-jfn58\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.669082 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-utilities\") pod \"community-operators-jfn58\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.669646 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-utilities\") pod \"community-operators-jfn58\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.669640 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-catalog-content\") pod \"community-operators-jfn58\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.696225 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7hjc\" (UniqueName: \"kubernetes.io/projected/30dea74f-070d-464e-8ed5-806cbd07b4f9-kube-api-access-j7hjc\") pod \"community-operators-jfn58\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:46 crc kubenswrapper[4871]: I1007 23:47:46.795951 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:47 crc kubenswrapper[4871]: I1007 23:47:47.489433 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jfn58"] Oct 07 23:47:47 crc kubenswrapper[4871]: I1007 23:47:47.699876 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jfn58" event={"ID":"30dea74f-070d-464e-8ed5-806cbd07b4f9","Type":"ContainerStarted","Data":"592e6f2015dc4bb628dc233639e724cf3276e64a536fc1ee95cce25c40e9623a"} Oct 07 23:47:48 crc kubenswrapper[4871]: I1007 23:47:48.712132 4871 generic.go:334] "Generic (PLEG): container finished" podID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerID="9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55" exitCode=0 Oct 07 23:47:48 crc kubenswrapper[4871]: I1007 23:47:48.712187 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jfn58" event={"ID":"30dea74f-070d-464e-8ed5-806cbd07b4f9","Type":"ContainerDied","Data":"9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55"} Oct 07 23:47:49 crc kubenswrapper[4871]: I1007 23:47:49.726756 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jfn58" event={"ID":"30dea74f-070d-464e-8ed5-806cbd07b4f9","Type":"ContainerStarted","Data":"66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f"} Oct 07 23:47:50 crc kubenswrapper[4871]: I1007 23:47:50.743221 4871 generic.go:334] "Generic (PLEG): container finished" podID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerID="66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f" exitCode=0 Oct 07 23:47:50 crc kubenswrapper[4871]: I1007 23:47:50.743734 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jfn58" event={"ID":"30dea74f-070d-464e-8ed5-806cbd07b4f9","Type":"ContainerDied","Data":"66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f"} Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.542712 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-w7c98"] Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.544197 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.546909 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.550770 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-qhpnc" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.552715 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-8dvm9"] Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.556037 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.560008 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-w7c98"] Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.570351 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-8dvm9"] Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625126 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-var-run\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625186 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-var-run-ovn\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625228 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-etc-ovs\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625282 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-var-log-ovn\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625300 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-scripts\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625381 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-var-lib\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625432 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5t6f\" (UniqueName: \"kubernetes.io/projected/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-kube-api-access-s5t6f\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625474 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-scripts\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625499 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25mg2\" (UniqueName: \"kubernetes.io/projected/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-kube-api-access-25mg2\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625522 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-var-run\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.625564 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-var-log\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727285 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-var-run\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727349 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-var-run-ovn\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727385 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-etc-ovs\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727429 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-var-log-ovn\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727446 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-scripts\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727664 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-var-run\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727688 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-var-run-ovn\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727701 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-var-log-ovn\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727694 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-etc-ovs\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.727913 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-var-lib\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.728021 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-var-lib\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.728051 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5t6f\" (UniqueName: \"kubernetes.io/projected/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-kube-api-access-s5t6f\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.728096 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-scripts\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.728113 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25mg2\" (UniqueName: \"kubernetes.io/projected/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-kube-api-access-25mg2\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.728131 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-var-run\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.728170 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-var-log\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.728249 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-var-run\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.728355 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-var-log\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.730040 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-scripts\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.730139 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-scripts\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.748601 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5t6f\" (UniqueName: \"kubernetes.io/projected/c38bb4f6-50ea-4b50-97f3-13a6838c4aae-kube-api-access-s5t6f\") pod \"ovn-controller-w7c98\" (UID: \"c38bb4f6-50ea-4b50-97f3-13a6838c4aae\") " pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.749644 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25mg2\" (UniqueName: \"kubernetes.io/projected/58f6bc11-48a7-48c8-a28b-5e02c8d284c5-kube-api-access-25mg2\") pod \"ovn-controller-ovs-8dvm9\" (UID: \"58f6bc11-48a7-48c8-a28b-5e02c8d284c5\") " pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.757603 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jfn58" event={"ID":"30dea74f-070d-464e-8ed5-806cbd07b4f9","Type":"ContainerStarted","Data":"e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8"} Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.778022 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jfn58" podStartSLOduration=3.234322289 podStartE2EDuration="5.778001551s" podCreationTimestamp="2025-10-07 23:47:46 +0000 UTC" firstStartedPulling="2025-10-07 23:47:48.714317981 +0000 UTC m=+5942.517016054" lastFinishedPulling="2025-10-07 23:47:51.257997233 +0000 UTC m=+5945.060695316" observedRunningTime="2025-10-07 23:47:51.776423049 +0000 UTC m=+5945.579121132" watchObservedRunningTime="2025-10-07 23:47:51.778001551 +0000 UTC m=+5945.580699624" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.862881 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w7c98" Oct 07 23:47:51 crc kubenswrapper[4871]: I1007 23:47:51.879067 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.450531 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-w7c98"] Oct 07 23:47:52 crc kubenswrapper[4871]: W1007 23:47:52.527127 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc38bb4f6_50ea_4b50_97f3_13a6838c4aae.slice/crio-ecc08981d050c443e56040e67ca25ab1fa342bd408ad158403e17a78c0906c58 WatchSource:0}: Error finding container ecc08981d050c443e56040e67ca25ab1fa342bd408ad158403e17a78c0906c58: Status 404 returned error can't find the container with id ecc08981d050c443e56040e67ca25ab1fa342bd408ad158403e17a78c0906c58 Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.713620 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-8dvm9"] Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.770334 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-8dvm9" event={"ID":"58f6bc11-48a7-48c8-a28b-5e02c8d284c5","Type":"ContainerStarted","Data":"d2a5a7879930c47eac2b77c9ad712c4b9dc3c6f54768d8c7aaafdff8e6cbfdf8"} Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.775985 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w7c98" event={"ID":"c38bb4f6-50ea-4b50-97f3-13a6838c4aae","Type":"ContainerStarted","Data":"ecc08981d050c443e56040e67ca25ab1fa342bd408ad158403e17a78c0906c58"} Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.813173 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-lcsmb"] Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.814634 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.824349 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.833478 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lcsmb"] Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.961064 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c553ffe-f347-4370-a2c5-c80ac5686859-config\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.961121 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2c553ffe-f347-4370-a2c5-c80ac5686859-ovs-rundir\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.961298 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ttpf\" (UniqueName: \"kubernetes.io/projected/2c553ffe-f347-4370-a2c5-c80ac5686859-kube-api-access-8ttpf\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:52 crc kubenswrapper[4871]: I1007 23:47:52.961373 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2c553ffe-f347-4370-a2c5-c80ac5686859-ovn-rundir\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.063237 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ttpf\" (UniqueName: \"kubernetes.io/projected/2c553ffe-f347-4370-a2c5-c80ac5686859-kube-api-access-8ttpf\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.063319 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2c553ffe-f347-4370-a2c5-c80ac5686859-ovn-rundir\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.063357 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c553ffe-f347-4370-a2c5-c80ac5686859-config\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.063388 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2c553ffe-f347-4370-a2c5-c80ac5686859-ovs-rundir\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.064253 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2c553ffe-f347-4370-a2c5-c80ac5686859-ovs-rundir\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.064256 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2c553ffe-f347-4370-a2c5-c80ac5686859-ovn-rundir\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.064945 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c553ffe-f347-4370-a2c5-c80ac5686859-config\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.081105 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ttpf\" (UniqueName: \"kubernetes.io/projected/2c553ffe-f347-4370-a2c5-c80ac5686859-kube-api-access-8ttpf\") pod \"ovn-controller-metrics-lcsmb\" (UID: \"2c553ffe-f347-4370-a2c5-c80ac5686859\") " pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.211466 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lcsmb" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.663961 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lcsmb"] Oct 07 23:47:53 crc kubenswrapper[4871]: W1007 23:47:53.666854 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c553ffe_f347_4370_a2c5_c80ac5686859.slice/crio-39df3bf8375d99a0b867a7de6d5cf2befb13a23cfda06dc641b50e3c25aba1d9 WatchSource:0}: Error finding container 39df3bf8375d99a0b867a7de6d5cf2befb13a23cfda06dc641b50e3c25aba1d9: Status 404 returned error can't find the container with id 39df3bf8375d99a0b867a7de6d5cf2befb13a23cfda06dc641b50e3c25aba1d9 Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.797180 4871 generic.go:334] "Generic (PLEG): container finished" podID="58f6bc11-48a7-48c8-a28b-5e02c8d284c5" containerID="02352ec37e91d9418cc4208065de3b093cb6394d798dc4f34c54cfefafaccb0a" exitCode=0 Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.797256 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-8dvm9" event={"ID":"58f6bc11-48a7-48c8-a28b-5e02c8d284c5","Type":"ContainerDied","Data":"02352ec37e91d9418cc4208065de3b093cb6394d798dc4f34c54cfefafaccb0a"} Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.799076 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lcsmb" event={"ID":"2c553ffe-f347-4370-a2c5-c80ac5686859","Type":"ContainerStarted","Data":"39df3bf8375d99a0b867a7de6d5cf2befb13a23cfda06dc641b50e3c25aba1d9"} Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.802413 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w7c98" event={"ID":"c38bb4f6-50ea-4b50-97f3-13a6838c4aae","Type":"ContainerStarted","Data":"c6c7701260480d3bebacaad7b4cf10f720a43e60ae87bd74cb40e0bbe7895479"} Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.802667 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-w7c98" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.852961 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-w7c98" podStartSLOduration=2.852943827 podStartE2EDuration="2.852943827s" podCreationTimestamp="2025-10-07 23:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:47:53.84015871 +0000 UTC m=+5947.642856803" watchObservedRunningTime="2025-10-07 23:47:53.852943827 +0000 UTC m=+5947.655641900" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.860656 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-gq7n9"] Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.862218 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-gq7n9" Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.868646 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-gq7n9"] Oct 07 23:47:53 crc kubenswrapper[4871]: I1007 23:47:53.983754 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s2kv\" (UniqueName: \"kubernetes.io/projected/3c9dee7b-aba0-4d5d-8562-0550d4db7584-kube-api-access-5s2kv\") pod \"octavia-db-create-gq7n9\" (UID: \"3c9dee7b-aba0-4d5d-8562-0550d4db7584\") " pod="openstack/octavia-db-create-gq7n9" Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.099103 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s2kv\" (UniqueName: \"kubernetes.io/projected/3c9dee7b-aba0-4d5d-8562-0550d4db7584-kube-api-access-5s2kv\") pod \"octavia-db-create-gq7n9\" (UID: \"3c9dee7b-aba0-4d5d-8562-0550d4db7584\") " pod="openstack/octavia-db-create-gq7n9" Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.139638 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s2kv\" (UniqueName: \"kubernetes.io/projected/3c9dee7b-aba0-4d5d-8562-0550d4db7584-kube-api-access-5s2kv\") pod \"octavia-db-create-gq7n9\" (UID: \"3c9dee7b-aba0-4d5d-8562-0550d4db7584\") " pod="openstack/octavia-db-create-gq7n9" Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.242664 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-gq7n9" Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.708209 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-gq7n9"] Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.812833 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lcsmb" event={"ID":"2c553ffe-f347-4370-a2c5-c80ac5686859","Type":"ContainerStarted","Data":"c976e6cfd71169fa2de914d9c4c27c120f6d954f026f7959dd32f4687860b4f3"} Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.814838 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-gq7n9" event={"ID":"3c9dee7b-aba0-4d5d-8562-0550d4db7584","Type":"ContainerStarted","Data":"80f88fd553f11bc454894d6299d66f67ec65eaec423a962fbd026b1d3dedc400"} Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.817506 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-8dvm9" event={"ID":"58f6bc11-48a7-48c8-a28b-5e02c8d284c5","Type":"ContainerStarted","Data":"012582fecdf5d61c70c88d94e3a057a743ce1cbb5293938762be292c22de15dd"} Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.817538 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-8dvm9" event={"ID":"58f6bc11-48a7-48c8-a28b-5e02c8d284c5","Type":"ContainerStarted","Data":"50e0a0676e772aab4b1e83bf11280c33a8c4e4b6f53c4ba9da48f714ca1de1e8"} Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.817565 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.817648 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.833430 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-lcsmb" podStartSLOduration=2.833404801 podStartE2EDuration="2.833404801s" podCreationTimestamp="2025-10-07 23:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:47:54.829631691 +0000 UTC m=+5948.632329774" watchObservedRunningTime="2025-10-07 23:47:54.833404801 +0000 UTC m=+5948.636102884" Oct 07 23:47:54 crc kubenswrapper[4871]: I1007 23:47:54.855444 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-8dvm9" podStartSLOduration=3.8554109519999997 podStartE2EDuration="3.855410952s" podCreationTimestamp="2025-10-07 23:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:47:54.853570333 +0000 UTC m=+5948.656268416" watchObservedRunningTime="2025-10-07 23:47:54.855410952 +0000 UTC m=+5948.658109035" Oct 07 23:47:55 crc kubenswrapper[4871]: I1007 23:47:55.026193 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:47:55 crc kubenswrapper[4871]: E1007 23:47:55.026425 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:47:55 crc kubenswrapper[4871]: I1007 23:47:55.833554 4871 generic.go:334] "Generic (PLEG): container finished" podID="3c9dee7b-aba0-4d5d-8562-0550d4db7584" containerID="a710dcd16e9ce1303317222eba462dce03c60048634e33123a848640a010055f" exitCode=0 Oct 07 23:47:55 crc kubenswrapper[4871]: I1007 23:47:55.835549 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-gq7n9" event={"ID":"3c9dee7b-aba0-4d5d-8562-0550d4db7584","Type":"ContainerDied","Data":"a710dcd16e9ce1303317222eba462dce03c60048634e33123a848640a010055f"} Oct 07 23:47:56 crc kubenswrapper[4871]: I1007 23:47:56.796235 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:56 crc kubenswrapper[4871]: I1007 23:47:56.796324 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:47:57 crc kubenswrapper[4871]: I1007 23:47:57.239998 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-gq7n9" Oct 07 23:47:57 crc kubenswrapper[4871]: I1007 23:47:57.377905 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s2kv\" (UniqueName: \"kubernetes.io/projected/3c9dee7b-aba0-4d5d-8562-0550d4db7584-kube-api-access-5s2kv\") pod \"3c9dee7b-aba0-4d5d-8562-0550d4db7584\" (UID: \"3c9dee7b-aba0-4d5d-8562-0550d4db7584\") " Oct 07 23:47:57 crc kubenswrapper[4871]: I1007 23:47:57.385960 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c9dee7b-aba0-4d5d-8562-0550d4db7584-kube-api-access-5s2kv" (OuterVolumeSpecName: "kube-api-access-5s2kv") pod "3c9dee7b-aba0-4d5d-8562-0550d4db7584" (UID: "3c9dee7b-aba0-4d5d-8562-0550d4db7584"). InnerVolumeSpecName "kube-api-access-5s2kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:47:57 crc kubenswrapper[4871]: I1007 23:47:57.480659 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s2kv\" (UniqueName: \"kubernetes.io/projected/3c9dee7b-aba0-4d5d-8562-0550d4db7584-kube-api-access-5s2kv\") on node \"crc\" DevicePath \"\"" Oct 07 23:47:57 crc kubenswrapper[4871]: I1007 23:47:57.862322 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-gq7n9" event={"ID":"3c9dee7b-aba0-4d5d-8562-0550d4db7584","Type":"ContainerDied","Data":"80f88fd553f11bc454894d6299d66f67ec65eaec423a962fbd026b1d3dedc400"} Oct 07 23:47:57 crc kubenswrapper[4871]: I1007 23:47:57.862357 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80f88fd553f11bc454894d6299d66f67ec65eaec423a962fbd026b1d3dedc400" Oct 07 23:47:57 crc kubenswrapper[4871]: I1007 23:47:57.862406 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-gq7n9" Oct 07 23:47:57 crc kubenswrapper[4871]: I1007 23:47:57.898439 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jfn58" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerName="registry-server" probeResult="failure" output=< Oct 07 23:47:57 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 23:47:57 crc kubenswrapper[4871]: > Oct 07 23:47:58 crc kubenswrapper[4871]: I1007 23:47:58.053979 4871 scope.go:117] "RemoveContainer" containerID="3b157e4a5da5c53c907cc20fdfbbab34137d9981a2ebbf608f65ba73fc44b28b" Oct 07 23:48:06 crc kubenswrapper[4871]: I1007 23:48:06.883618 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:48:06 crc kubenswrapper[4871]: I1007 23:48:06.893491 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-8bf7-account-create-s55bk"] Oct 07 23:48:06 crc kubenswrapper[4871]: E1007 23:48:06.894199 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c9dee7b-aba0-4d5d-8562-0550d4db7584" containerName="mariadb-database-create" Oct 07 23:48:06 crc kubenswrapper[4871]: I1007 23:48:06.894219 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c9dee7b-aba0-4d5d-8562-0550d4db7584" containerName="mariadb-database-create" Oct 07 23:48:06 crc kubenswrapper[4871]: I1007 23:48:06.894477 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c9dee7b-aba0-4d5d-8562-0550d4db7584" containerName="mariadb-database-create" Oct 07 23:48:06 crc kubenswrapper[4871]: I1007 23:48:06.895471 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-8bf7-account-create-s55bk" Oct 07 23:48:06 crc kubenswrapper[4871]: I1007 23:48:06.897432 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Oct 07 23:48:06 crc kubenswrapper[4871]: I1007 23:48:06.913569 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-8bf7-account-create-s55bk"] Oct 07 23:48:06 crc kubenswrapper[4871]: I1007 23:48:06.939282 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.002966 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp7h2\" (UniqueName: \"kubernetes.io/projected/a00b25ff-46ef-4d87-ad93-81543f7851fa-kube-api-access-lp7h2\") pod \"octavia-8bf7-account-create-s55bk\" (UID: \"a00b25ff-46ef-4d87-ad93-81543f7851fa\") " pod="openstack/octavia-8bf7-account-create-s55bk" Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.105962 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp7h2\" (UniqueName: \"kubernetes.io/projected/a00b25ff-46ef-4d87-ad93-81543f7851fa-kube-api-access-lp7h2\") pod \"octavia-8bf7-account-create-s55bk\" (UID: \"a00b25ff-46ef-4d87-ad93-81543f7851fa\") " pod="openstack/octavia-8bf7-account-create-s55bk" Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.119418 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jfn58"] Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.130447 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp7h2\" (UniqueName: \"kubernetes.io/projected/a00b25ff-46ef-4d87-ad93-81543f7851fa-kube-api-access-lp7h2\") pod \"octavia-8bf7-account-create-s55bk\" (UID: \"a00b25ff-46ef-4d87-ad93-81543f7851fa\") " pod="openstack/octavia-8bf7-account-create-s55bk" Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.216348 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-8bf7-account-create-s55bk" Oct 07 23:48:07 crc kubenswrapper[4871]: W1007 23:48:07.707748 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda00b25ff_46ef_4d87_ad93_81543f7851fa.slice/crio-509e644e406612afe8c720dce21c70e1e0aeb240ae008a0bce030a008859254c WatchSource:0}: Error finding container 509e644e406612afe8c720dce21c70e1e0aeb240ae008a0bce030a008859254c: Status 404 returned error can't find the container with id 509e644e406612afe8c720dce21c70e1e0aeb240ae008a0bce030a008859254c Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.709265 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-8bf7-account-create-s55bk"] Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.981983 4871 generic.go:334] "Generic (PLEG): container finished" podID="a00b25ff-46ef-4d87-ad93-81543f7851fa" containerID="f14c14f5ade2edcf73b172090cb7b45d53273b48c7614eb77f60a4474337ab0d" exitCode=0 Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.982121 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-8bf7-account-create-s55bk" event={"ID":"a00b25ff-46ef-4d87-ad93-81543f7851fa","Type":"ContainerDied","Data":"f14c14f5ade2edcf73b172090cb7b45d53273b48c7614eb77f60a4474337ab0d"} Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.983378 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-8bf7-account-create-s55bk" event={"ID":"a00b25ff-46ef-4d87-ad93-81543f7851fa","Type":"ContainerStarted","Data":"509e644e406612afe8c720dce21c70e1e0aeb240ae008a0bce030a008859254c"} Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.982574 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:48:07 crc kubenswrapper[4871]: I1007 23:48:07.983478 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jfn58" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerName="registry-server" containerID="cri-o://e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8" gracePeriod=2 Oct 07 23:48:07 crc kubenswrapper[4871]: E1007 23:48:07.984120 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.504627 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.636363 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-catalog-content\") pod \"30dea74f-070d-464e-8ed5-806cbd07b4f9\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.636428 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-utilities\") pod \"30dea74f-070d-464e-8ed5-806cbd07b4f9\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.636540 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7hjc\" (UniqueName: \"kubernetes.io/projected/30dea74f-070d-464e-8ed5-806cbd07b4f9-kube-api-access-j7hjc\") pod \"30dea74f-070d-464e-8ed5-806cbd07b4f9\" (UID: \"30dea74f-070d-464e-8ed5-806cbd07b4f9\") " Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.638602 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-utilities" (OuterVolumeSpecName: "utilities") pod "30dea74f-070d-464e-8ed5-806cbd07b4f9" (UID: "30dea74f-070d-464e-8ed5-806cbd07b4f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.642197 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30dea74f-070d-464e-8ed5-806cbd07b4f9-kube-api-access-j7hjc" (OuterVolumeSpecName: "kube-api-access-j7hjc") pod "30dea74f-070d-464e-8ed5-806cbd07b4f9" (UID: "30dea74f-070d-464e-8ed5-806cbd07b4f9"). InnerVolumeSpecName "kube-api-access-j7hjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.695450 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30dea74f-070d-464e-8ed5-806cbd07b4f9" (UID: "30dea74f-070d-464e-8ed5-806cbd07b4f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.739606 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.739670 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30dea74f-070d-464e-8ed5-806cbd07b4f9-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.739689 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7hjc\" (UniqueName: \"kubernetes.io/projected/30dea74f-070d-464e-8ed5-806cbd07b4f9-kube-api-access-j7hjc\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.997559 4871 generic.go:334] "Generic (PLEG): container finished" podID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerID="e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8" exitCode=0 Oct 07 23:48:08 crc kubenswrapper[4871]: I1007 23:48:08.997717 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jfn58" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.000914 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jfn58" event={"ID":"30dea74f-070d-464e-8ed5-806cbd07b4f9","Type":"ContainerDied","Data":"e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8"} Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.000984 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jfn58" event={"ID":"30dea74f-070d-464e-8ed5-806cbd07b4f9","Type":"ContainerDied","Data":"592e6f2015dc4bb628dc233639e724cf3276e64a536fc1ee95cce25c40e9623a"} Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.001018 4871 scope.go:117] "RemoveContainer" containerID="e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.033109 4871 scope.go:117] "RemoveContainer" containerID="66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.048581 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jfn58"] Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.058779 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jfn58"] Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.077951 4871 scope.go:117] "RemoveContainer" containerID="9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.118473 4871 scope.go:117] "RemoveContainer" containerID="e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8" Oct 07 23:48:09 crc kubenswrapper[4871]: E1007 23:48:09.119216 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8\": container with ID starting with e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8 not found: ID does not exist" containerID="e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.119266 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8"} err="failed to get container status \"e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8\": rpc error: code = NotFound desc = could not find container \"e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8\": container with ID starting with e3f0e196ad90f8856d815a5929887b14fe198ce6a17e94ce55c03bc76c69a6f8 not found: ID does not exist" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.119433 4871 scope.go:117] "RemoveContainer" containerID="66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f" Oct 07 23:48:09 crc kubenswrapper[4871]: E1007 23:48:09.119820 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f\": container with ID starting with 66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f not found: ID does not exist" containerID="66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.119859 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f"} err="failed to get container status \"66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f\": rpc error: code = NotFound desc = could not find container \"66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f\": container with ID starting with 66450e3f3e9013e37d54ed0037a8c25df76514133f0699fce049a3b89f819a0f not found: ID does not exist" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.119886 4871 scope.go:117] "RemoveContainer" containerID="9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55" Oct 07 23:48:09 crc kubenswrapper[4871]: E1007 23:48:09.120206 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55\": container with ID starting with 9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55 not found: ID does not exist" containerID="9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.120231 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55"} err="failed to get container status \"9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55\": rpc error: code = NotFound desc = could not find container \"9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55\": container with ID starting with 9997ad03bd2b16fed7373170f4c115385f4993f38f8ba9d60f716755a6ed0a55 not found: ID does not exist" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.442992 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-8bf7-account-create-s55bk" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.553941 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp7h2\" (UniqueName: \"kubernetes.io/projected/a00b25ff-46ef-4d87-ad93-81543f7851fa-kube-api-access-lp7h2\") pod \"a00b25ff-46ef-4d87-ad93-81543f7851fa\" (UID: \"a00b25ff-46ef-4d87-ad93-81543f7851fa\") " Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.566617 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00b25ff-46ef-4d87-ad93-81543f7851fa-kube-api-access-lp7h2" (OuterVolumeSpecName: "kube-api-access-lp7h2") pod "a00b25ff-46ef-4d87-ad93-81543f7851fa" (UID: "a00b25ff-46ef-4d87-ad93-81543f7851fa"). InnerVolumeSpecName "kube-api-access-lp7h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:48:09 crc kubenswrapper[4871]: I1007 23:48:09.656436 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp7h2\" (UniqueName: \"kubernetes.io/projected/a00b25ff-46ef-4d87-ad93-81543f7851fa-kube-api-access-lp7h2\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:10 crc kubenswrapper[4871]: I1007 23:48:10.012981 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-8bf7-account-create-s55bk" Oct 07 23:48:10 crc kubenswrapper[4871]: I1007 23:48:10.013041 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-8bf7-account-create-s55bk" event={"ID":"a00b25ff-46ef-4d87-ad93-81543f7851fa","Type":"ContainerDied","Data":"509e644e406612afe8c720dce21c70e1e0aeb240ae008a0bce030a008859254c"} Oct 07 23:48:10 crc kubenswrapper[4871]: I1007 23:48:10.013090 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="509e644e406612afe8c720dce21c70e1e0aeb240ae008a0bce030a008859254c" Oct 07 23:48:11 crc kubenswrapper[4871]: I1007 23:48:11.005406 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" path="/var/lib/kubelet/pods/30dea74f-070d-464e-8ed5-806cbd07b4f9/volumes" Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.905179 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-x4fxr"] Oct 07 23:48:13 crc kubenswrapper[4871]: E1007 23:48:13.906026 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00b25ff-46ef-4d87-ad93-81543f7851fa" containerName="mariadb-account-create" Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.906043 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00b25ff-46ef-4d87-ad93-81543f7851fa" containerName="mariadb-account-create" Oct 07 23:48:13 crc kubenswrapper[4871]: E1007 23:48:13.906060 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerName="extract-utilities" Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.906071 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerName="extract-utilities" Oct 07 23:48:13 crc kubenswrapper[4871]: E1007 23:48:13.906082 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerName="extract-content" Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.906090 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerName="extract-content" Oct 07 23:48:13 crc kubenswrapper[4871]: E1007 23:48:13.906124 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerName="registry-server" Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.906132 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerName="registry-server" Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.906360 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00b25ff-46ef-4d87-ad93-81543f7851fa" containerName="mariadb-account-create" Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.906386 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="30dea74f-070d-464e-8ed5-806cbd07b4f9" containerName="registry-server" Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.907157 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-x4fxr" Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.922123 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-x4fxr"] Oct 07 23:48:13 crc kubenswrapper[4871]: I1007 23:48:13.966753 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8qs2\" (UniqueName: \"kubernetes.io/projected/555b619b-6b5e-403c-ab31-4484710488a3-kube-api-access-h8qs2\") pod \"octavia-persistence-db-create-x4fxr\" (UID: \"555b619b-6b5e-403c-ab31-4484710488a3\") " pod="openstack/octavia-persistence-db-create-x4fxr" Oct 07 23:48:14 crc kubenswrapper[4871]: I1007 23:48:14.069177 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8qs2\" (UniqueName: \"kubernetes.io/projected/555b619b-6b5e-403c-ab31-4484710488a3-kube-api-access-h8qs2\") pod \"octavia-persistence-db-create-x4fxr\" (UID: \"555b619b-6b5e-403c-ab31-4484710488a3\") " pod="openstack/octavia-persistence-db-create-x4fxr" Oct 07 23:48:14 crc kubenswrapper[4871]: I1007 23:48:14.097378 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8qs2\" (UniqueName: \"kubernetes.io/projected/555b619b-6b5e-403c-ab31-4484710488a3-kube-api-access-h8qs2\") pod \"octavia-persistence-db-create-x4fxr\" (UID: \"555b619b-6b5e-403c-ab31-4484710488a3\") " pod="openstack/octavia-persistence-db-create-x4fxr" Oct 07 23:48:14 crc kubenswrapper[4871]: I1007 23:48:14.225068 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-x4fxr" Oct 07 23:48:14 crc kubenswrapper[4871]: W1007 23:48:14.718471 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod555b619b_6b5e_403c_ab31_4484710488a3.slice/crio-41f98b3a076ab9bf46f15b49c334cfc45c29c94a207887a3c25838087afc606d WatchSource:0}: Error finding container 41f98b3a076ab9bf46f15b49c334cfc45c29c94a207887a3c25838087afc606d: Status 404 returned error can't find the container with id 41f98b3a076ab9bf46f15b49c334cfc45c29c94a207887a3c25838087afc606d Oct 07 23:48:14 crc kubenswrapper[4871]: I1007 23:48:14.723169 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-x4fxr"] Oct 07 23:48:15 crc kubenswrapper[4871]: I1007 23:48:15.066371 4871 generic.go:334] "Generic (PLEG): container finished" podID="555b619b-6b5e-403c-ab31-4484710488a3" containerID="b06488eec2fb7f1159bede2eb9d48b3ade5e20c2bc3d2f32a86e5f344eb5facb" exitCode=0 Oct 07 23:48:15 crc kubenswrapper[4871]: I1007 23:48:15.066453 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-x4fxr" event={"ID":"555b619b-6b5e-403c-ab31-4484710488a3","Type":"ContainerDied","Data":"b06488eec2fb7f1159bede2eb9d48b3ade5e20c2bc3d2f32a86e5f344eb5facb"} Oct 07 23:48:15 crc kubenswrapper[4871]: I1007 23:48:15.066602 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-x4fxr" event={"ID":"555b619b-6b5e-403c-ab31-4484710488a3","Type":"ContainerStarted","Data":"41f98b3a076ab9bf46f15b49c334cfc45c29c94a207887a3c25838087afc606d"} Oct 07 23:48:16 crc kubenswrapper[4871]: I1007 23:48:16.384010 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-x4fxr" Oct 07 23:48:16 crc kubenswrapper[4871]: I1007 23:48:16.417910 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8qs2\" (UniqueName: \"kubernetes.io/projected/555b619b-6b5e-403c-ab31-4484710488a3-kube-api-access-h8qs2\") pod \"555b619b-6b5e-403c-ab31-4484710488a3\" (UID: \"555b619b-6b5e-403c-ab31-4484710488a3\") " Oct 07 23:48:16 crc kubenswrapper[4871]: I1007 23:48:16.423690 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/555b619b-6b5e-403c-ab31-4484710488a3-kube-api-access-h8qs2" (OuterVolumeSpecName: "kube-api-access-h8qs2") pod "555b619b-6b5e-403c-ab31-4484710488a3" (UID: "555b619b-6b5e-403c-ab31-4484710488a3"). InnerVolumeSpecName "kube-api-access-h8qs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:48:16 crc kubenswrapper[4871]: I1007 23:48:16.521122 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8qs2\" (UniqueName: \"kubernetes.io/projected/555b619b-6b5e-403c-ab31-4484710488a3-kube-api-access-h8qs2\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:17 crc kubenswrapper[4871]: I1007 23:48:17.087166 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-x4fxr" event={"ID":"555b619b-6b5e-403c-ab31-4484710488a3","Type":"ContainerDied","Data":"41f98b3a076ab9bf46f15b49c334cfc45c29c94a207887a3c25838087afc606d"} Oct 07 23:48:17 crc kubenswrapper[4871]: I1007 23:48:17.087537 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41f98b3a076ab9bf46f15b49c334cfc45c29c94a207887a3c25838087afc606d" Oct 07 23:48:17 crc kubenswrapper[4871]: I1007 23:48:17.087201 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-x4fxr" Oct 07 23:48:19 crc kubenswrapper[4871]: I1007 23:48:19.982776 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:48:19 crc kubenswrapper[4871]: E1007 23:48:19.984785 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.015950 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-f4cd-account-create-rdx2p"] Oct 07 23:48:25 crc kubenswrapper[4871]: E1007 23:48:25.016860 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="555b619b-6b5e-403c-ab31-4484710488a3" containerName="mariadb-database-create" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.016874 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="555b619b-6b5e-403c-ab31-4484710488a3" containerName="mariadb-database-create" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.017184 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="555b619b-6b5e-403c-ab31-4484710488a3" containerName="mariadb-database-create" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.018020 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f4cd-account-create-rdx2p" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.024452 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.026869 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-f4cd-account-create-rdx2p"] Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.118359 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjw87\" (UniqueName: \"kubernetes.io/projected/dbe71c83-937f-4cce-8ea2-26c78aeb84de-kube-api-access-wjw87\") pod \"octavia-f4cd-account-create-rdx2p\" (UID: \"dbe71c83-937f-4cce-8ea2-26c78aeb84de\") " pod="openstack/octavia-f4cd-account-create-rdx2p" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.221979 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjw87\" (UniqueName: \"kubernetes.io/projected/dbe71c83-937f-4cce-8ea2-26c78aeb84de-kube-api-access-wjw87\") pod \"octavia-f4cd-account-create-rdx2p\" (UID: \"dbe71c83-937f-4cce-8ea2-26c78aeb84de\") " pod="openstack/octavia-f4cd-account-create-rdx2p" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.246578 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjw87\" (UniqueName: \"kubernetes.io/projected/dbe71c83-937f-4cce-8ea2-26c78aeb84de-kube-api-access-wjw87\") pod \"octavia-f4cd-account-create-rdx2p\" (UID: \"dbe71c83-937f-4cce-8ea2-26c78aeb84de\") " pod="openstack/octavia-f4cd-account-create-rdx2p" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.346516 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f4cd-account-create-rdx2p" Oct 07 23:48:25 crc kubenswrapper[4871]: I1007 23:48:25.811312 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-f4cd-account-create-rdx2p"] Oct 07 23:48:26 crc kubenswrapper[4871]: I1007 23:48:26.174014 4871 generic.go:334] "Generic (PLEG): container finished" podID="dbe71c83-937f-4cce-8ea2-26c78aeb84de" containerID="a1b1e2021025c96be9d4900631fc35284c4709777780086d38898e33b6b85fb1" exitCode=0 Oct 07 23:48:26 crc kubenswrapper[4871]: I1007 23:48:26.174125 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f4cd-account-create-rdx2p" event={"ID":"dbe71c83-937f-4cce-8ea2-26c78aeb84de","Type":"ContainerDied","Data":"a1b1e2021025c96be9d4900631fc35284c4709777780086d38898e33b6b85fb1"} Oct 07 23:48:26 crc kubenswrapper[4871]: I1007 23:48:26.174369 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f4cd-account-create-rdx2p" event={"ID":"dbe71c83-937f-4cce-8ea2-26c78aeb84de","Type":"ContainerStarted","Data":"e0c0e2926f9fe9a6bbe9dab33289bc1c234fd6e40ca3a638000f8886c6726e77"} Oct 07 23:48:26 crc kubenswrapper[4871]: I1007 23:48:26.910554 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-w7c98" podUID="c38bb4f6-50ea-4b50-97f3-13a6838c4aae" containerName="ovn-controller" probeResult="failure" output=< Oct 07 23:48:26 crc kubenswrapper[4871]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 07 23:48:26 crc kubenswrapper[4871]: > Oct 07 23:48:26 crc kubenswrapper[4871]: I1007 23:48:26.929568 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:48:26 crc kubenswrapper[4871]: I1007 23:48:26.931084 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-8dvm9" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.059246 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-w7c98-config-p788f"] Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.061047 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.063490 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.070059 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-w7c98-config-p788f"] Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.165051 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.165159 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-log-ovn\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.165217 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwqxt\" (UniqueName: \"kubernetes.io/projected/9251a28e-1411-44f1-9fec-44d03b549694-kube-api-access-vwqxt\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.165290 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-additional-scripts\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.165359 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-scripts\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.165547 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run-ovn\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.267496 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run-ovn\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.267942 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run-ovn\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.267973 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.268019 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.268023 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-log-ovn\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.268065 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-log-ovn\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.268111 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwqxt\" (UniqueName: \"kubernetes.io/projected/9251a28e-1411-44f1-9fec-44d03b549694-kube-api-access-vwqxt\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.268153 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-additional-scripts\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.268186 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-scripts\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.269120 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-additional-scripts\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.271359 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-scripts\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.287207 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwqxt\" (UniqueName: \"kubernetes.io/projected/9251a28e-1411-44f1-9fec-44d03b549694-kube-api-access-vwqxt\") pod \"ovn-controller-w7c98-config-p788f\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.385635 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.518562 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f4cd-account-create-rdx2p" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.675076 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjw87\" (UniqueName: \"kubernetes.io/projected/dbe71c83-937f-4cce-8ea2-26c78aeb84de-kube-api-access-wjw87\") pod \"dbe71c83-937f-4cce-8ea2-26c78aeb84de\" (UID: \"dbe71c83-937f-4cce-8ea2-26c78aeb84de\") " Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.690757 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbe71c83-937f-4cce-8ea2-26c78aeb84de-kube-api-access-wjw87" (OuterVolumeSpecName: "kube-api-access-wjw87") pod "dbe71c83-937f-4cce-8ea2-26c78aeb84de" (UID: "dbe71c83-937f-4cce-8ea2-26c78aeb84de"). InnerVolumeSpecName "kube-api-access-wjw87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.777348 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjw87\" (UniqueName: \"kubernetes.io/projected/dbe71c83-937f-4cce-8ea2-26c78aeb84de-kube-api-access-wjw87\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:27 crc kubenswrapper[4871]: I1007 23:48:27.812387 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-w7c98-config-p788f"] Oct 07 23:48:27 crc kubenswrapper[4871]: W1007 23:48:27.813876 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9251a28e_1411_44f1_9fec_44d03b549694.slice/crio-b984514d59faca315fb5e4d791c601550f46914b7005f2e8b5b2303a45d3044d WatchSource:0}: Error finding container b984514d59faca315fb5e4d791c601550f46914b7005f2e8b5b2303a45d3044d: Status 404 returned error can't find the container with id b984514d59faca315fb5e4d791c601550f46914b7005f2e8b5b2303a45d3044d Oct 07 23:48:28 crc kubenswrapper[4871]: I1007 23:48:28.196259 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f4cd-account-create-rdx2p" event={"ID":"dbe71c83-937f-4cce-8ea2-26c78aeb84de","Type":"ContainerDied","Data":"e0c0e2926f9fe9a6bbe9dab33289bc1c234fd6e40ca3a638000f8886c6726e77"} Oct 07 23:48:28 crc kubenswrapper[4871]: I1007 23:48:28.196527 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0c0e2926f9fe9a6bbe9dab33289bc1c234fd6e40ca3a638000f8886c6726e77" Oct 07 23:48:28 crc kubenswrapper[4871]: I1007 23:48:28.196598 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f4cd-account-create-rdx2p" Oct 07 23:48:28 crc kubenswrapper[4871]: I1007 23:48:28.201487 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w7c98-config-p788f" event={"ID":"9251a28e-1411-44f1-9fec-44d03b549694","Type":"ContainerStarted","Data":"57214d8df072bd58d6bd487bbb798e194944cf9f95d2de5da11d404777dbe252"} Oct 07 23:48:28 crc kubenswrapper[4871]: I1007 23:48:28.201826 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w7c98-config-p788f" event={"ID":"9251a28e-1411-44f1-9fec-44d03b549694","Type":"ContainerStarted","Data":"b984514d59faca315fb5e4d791c601550f46914b7005f2e8b5b2303a45d3044d"} Oct 07 23:48:28 crc kubenswrapper[4871]: I1007 23:48:28.225340 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-w7c98-config-p788f" podStartSLOduration=1.225325963 podStartE2EDuration="1.225325963s" podCreationTimestamp="2025-10-07 23:48:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:48:28.22176517 +0000 UTC m=+5982.024463243" watchObservedRunningTime="2025-10-07 23:48:28.225325963 +0000 UTC m=+5982.028024036" Oct 07 23:48:29 crc kubenswrapper[4871]: I1007 23:48:29.214044 4871 generic.go:334] "Generic (PLEG): container finished" podID="9251a28e-1411-44f1-9fec-44d03b549694" containerID="57214d8df072bd58d6bd487bbb798e194944cf9f95d2de5da11d404777dbe252" exitCode=0 Oct 07 23:48:29 crc kubenswrapper[4871]: I1007 23:48:29.214157 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w7c98-config-p788f" event={"ID":"9251a28e-1411-44f1-9fec-44d03b549694","Type":"ContainerDied","Data":"57214d8df072bd58d6bd487bbb798e194944cf9f95d2de5da11d404777dbe252"} Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.649594 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.761932 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwqxt\" (UniqueName: \"kubernetes.io/projected/9251a28e-1411-44f1-9fec-44d03b549694-kube-api-access-vwqxt\") pod \"9251a28e-1411-44f1-9fec-44d03b549694\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.762069 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-scripts\") pod \"9251a28e-1411-44f1-9fec-44d03b549694\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.762241 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run\") pod \"9251a28e-1411-44f1-9fec-44d03b549694\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.762287 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run-ovn\") pod \"9251a28e-1411-44f1-9fec-44d03b549694\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.762347 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-additional-scripts\") pod \"9251a28e-1411-44f1-9fec-44d03b549694\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.762382 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-log-ovn\") pod \"9251a28e-1411-44f1-9fec-44d03b549694\" (UID: \"9251a28e-1411-44f1-9fec-44d03b549694\") " Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.762396 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9251a28e-1411-44f1-9fec-44d03b549694" (UID: "9251a28e-1411-44f1-9fec-44d03b549694"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.762811 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9251a28e-1411-44f1-9fec-44d03b549694" (UID: "9251a28e-1411-44f1-9fec-44d03b549694"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.763071 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9251a28e-1411-44f1-9fec-44d03b549694" (UID: "9251a28e-1411-44f1-9fec-44d03b549694"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.763314 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-scripts" (OuterVolumeSpecName: "scripts") pod "9251a28e-1411-44f1-9fec-44d03b549694" (UID: "9251a28e-1411-44f1-9fec-44d03b549694"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.763336 4871 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.763356 4871 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.763371 4871 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.763409 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run" (OuterVolumeSpecName: "var-run") pod "9251a28e-1411-44f1-9fec-44d03b549694" (UID: "9251a28e-1411-44f1-9fec-44d03b549694"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.769799 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9251a28e-1411-44f1-9fec-44d03b549694-kube-api-access-vwqxt" (OuterVolumeSpecName: "kube-api-access-vwqxt") pod "9251a28e-1411-44f1-9fec-44d03b549694" (UID: "9251a28e-1411-44f1-9fec-44d03b549694"). InnerVolumeSpecName "kube-api-access-vwqxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.864555 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwqxt\" (UniqueName: \"kubernetes.io/projected/9251a28e-1411-44f1-9fec-44d03b549694-kube-api-access-vwqxt\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.864933 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9251a28e-1411-44f1-9fec-44d03b549694-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:30 crc kubenswrapper[4871]: I1007 23:48:30.864942 4871 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9251a28e-1411-44f1-9fec-44d03b549694-var-run\") on node \"crc\" DevicePath \"\"" Oct 07 23:48:31 crc kubenswrapper[4871]: I1007 23:48:31.238143 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-w7c98-config-p788f" event={"ID":"9251a28e-1411-44f1-9fec-44d03b549694","Type":"ContainerDied","Data":"b984514d59faca315fb5e4d791c601550f46914b7005f2e8b5b2303a45d3044d"} Oct 07 23:48:31 crc kubenswrapper[4871]: I1007 23:48:31.238212 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-w7c98-config-p788f" Oct 07 23:48:31 crc kubenswrapper[4871]: I1007 23:48:31.238222 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b984514d59faca315fb5e4d791c601550f46914b7005f2e8b5b2303a45d3044d" Oct 07 23:48:31 crc kubenswrapper[4871]: I1007 23:48:31.322893 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-w7c98-config-p788f"] Oct 07 23:48:31 crc kubenswrapper[4871]: I1007 23:48:31.335692 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-w7c98-config-p788f"] Oct 07 23:48:31 crc kubenswrapper[4871]: I1007 23:48:31.953268 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-w7c98" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.018730 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-7bb5c546c4-ms9lr"] Oct 07 23:48:32 crc kubenswrapper[4871]: E1007 23:48:32.019347 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe71c83-937f-4cce-8ea2-26c78aeb84de" containerName="mariadb-account-create" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.019372 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe71c83-937f-4cce-8ea2-26c78aeb84de" containerName="mariadb-account-create" Oct 07 23:48:32 crc kubenswrapper[4871]: E1007 23:48:32.019423 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9251a28e-1411-44f1-9fec-44d03b549694" containerName="ovn-config" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.019433 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9251a28e-1411-44f1-9fec-44d03b549694" containerName="ovn-config" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.019666 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9251a28e-1411-44f1-9fec-44d03b549694" containerName="ovn-config" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.019695 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe71c83-937f-4cce-8ea2-26c78aeb84de" containerName="mariadb-account-create" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.022339 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.026409 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.026619 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.035781 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-7bb5c546c4-ms9lr"] Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.048438 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-9zlhz" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.195138 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/446688c4-a538-417b-97d5-a8c2b9203f3c-octavia-run\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.195762 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446688c4-a538-417b-97d5-a8c2b9203f3c-config-data\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.195971 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/446688c4-a538-417b-97d5-a8c2b9203f3c-config-data-merged\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.196097 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446688c4-a538-417b-97d5-a8c2b9203f3c-scripts\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.196185 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446688c4-a538-417b-97d5-a8c2b9203f3c-combined-ca-bundle\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.297646 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446688c4-a538-417b-97d5-a8c2b9203f3c-scripts\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.297709 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446688c4-a538-417b-97d5-a8c2b9203f3c-combined-ca-bundle\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.297774 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/446688c4-a538-417b-97d5-a8c2b9203f3c-octavia-run\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.297829 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446688c4-a538-417b-97d5-a8c2b9203f3c-config-data\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.297873 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/446688c4-a538-417b-97d5-a8c2b9203f3c-config-data-merged\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.298286 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/446688c4-a538-417b-97d5-a8c2b9203f3c-config-data-merged\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.298902 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/446688c4-a538-417b-97d5-a8c2b9203f3c-octavia-run\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.303286 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/446688c4-a538-417b-97d5-a8c2b9203f3c-scripts\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.303858 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446688c4-a538-417b-97d5-a8c2b9203f3c-config-data\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.310566 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446688c4-a538-417b-97d5-a8c2b9203f3c-combined-ca-bundle\") pod \"octavia-api-7bb5c546c4-ms9lr\" (UID: \"446688c4-a538-417b-97d5-a8c2b9203f3c\") " pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.359634 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.834884 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-7bb5c546c4-ms9lr"] Oct 07 23:48:32 crc kubenswrapper[4871]: W1007 23:48:32.836500 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod446688c4_a538_417b_97d5_a8c2b9203f3c.slice/crio-f96d5e8fa9bd32c8d36c81ec759d2e4e8414d47851e454712358c29374ba7dd3 WatchSource:0}: Error finding container f96d5e8fa9bd32c8d36c81ec759d2e4e8414d47851e454712358c29374ba7dd3: Status 404 returned error can't find the container with id f96d5e8fa9bd32c8d36c81ec759d2e4e8414d47851e454712358c29374ba7dd3 Oct 07 23:48:32 crc kubenswrapper[4871]: I1007 23:48:32.995670 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9251a28e-1411-44f1-9fec-44d03b549694" path="/var/lib/kubelet/pods/9251a28e-1411-44f1-9fec-44d03b549694/volumes" Oct 07 23:48:33 crc kubenswrapper[4871]: I1007 23:48:33.253553 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7bb5c546c4-ms9lr" event={"ID":"446688c4-a538-417b-97d5-a8c2b9203f3c","Type":"ContainerStarted","Data":"f96d5e8fa9bd32c8d36c81ec759d2e4e8414d47851e454712358c29374ba7dd3"} Oct 07 23:48:34 crc kubenswrapper[4871]: I1007 23:48:34.982406 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:48:34 crc kubenswrapper[4871]: E1007 23:48:34.982777 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:48:43 crc kubenswrapper[4871]: I1007 23:48:43.350838 4871 generic.go:334] "Generic (PLEG): container finished" podID="446688c4-a538-417b-97d5-a8c2b9203f3c" containerID="4e2f6c901c49760cadbbfdcd7a9052c3553b1795f77c7b37bfcea0b590fceabc" exitCode=0 Oct 07 23:48:43 crc kubenswrapper[4871]: I1007 23:48:43.351122 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7bb5c546c4-ms9lr" event={"ID":"446688c4-a538-417b-97d5-a8c2b9203f3c","Type":"ContainerDied","Data":"4e2f6c901c49760cadbbfdcd7a9052c3553b1795f77c7b37bfcea0b590fceabc"} Oct 07 23:48:44 crc kubenswrapper[4871]: I1007 23:48:44.360872 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7bb5c546c4-ms9lr" event={"ID":"446688c4-a538-417b-97d5-a8c2b9203f3c","Type":"ContainerStarted","Data":"62471dadf65643e8f7de056f8fdeea7237516d3322c904446a1c0e1227b73418"} Oct 07 23:48:44 crc kubenswrapper[4871]: I1007 23:48:44.361211 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7bb5c546c4-ms9lr" event={"ID":"446688c4-a538-417b-97d5-a8c2b9203f3c","Type":"ContainerStarted","Data":"820196817d291dfd94e24c8085a94a38fbf2bae08f4763a1c771d79c0160953b"} Oct 07 23:48:44 crc kubenswrapper[4871]: I1007 23:48:44.361334 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:44 crc kubenswrapper[4871]: I1007 23:48:44.386438 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-7bb5c546c4-ms9lr" podStartSLOduration=3.415261964 podStartE2EDuration="13.386415185s" podCreationTimestamp="2025-10-07 23:48:31 +0000 UTC" firstStartedPulling="2025-10-07 23:48:32.839131684 +0000 UTC m=+5986.641829757" lastFinishedPulling="2025-10-07 23:48:42.810284905 +0000 UTC m=+5996.612982978" observedRunningTime="2025-10-07 23:48:44.377655074 +0000 UTC m=+5998.180353137" watchObservedRunningTime="2025-10-07 23:48:44.386415185 +0000 UTC m=+5998.189113248" Oct 07 23:48:45 crc kubenswrapper[4871]: I1007 23:48:45.370046 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:48:47 crc kubenswrapper[4871]: I1007 23:48:47.982712 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:48:48 crc kubenswrapper[4871]: I1007 23:48:48.400382 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"cee94e818dbd839f5c342b9e271035e3e73be566c36380410bb63a203783329e"} Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.610325 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-ckzbv"] Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.612847 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.614856 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.615463 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.615522 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.627124 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-ckzbv"] Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.690276 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13bcfd35-b654-40e4-88d8-fc290ba96b14-scripts\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.690478 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/13bcfd35-b654-40e4-88d8-fc290ba96b14-hm-ports\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.690514 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/13bcfd35-b654-40e4-88d8-fc290ba96b14-config-data-merged\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.690539 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13bcfd35-b654-40e4-88d8-fc290ba96b14-config-data\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.791974 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13bcfd35-b654-40e4-88d8-fc290ba96b14-scripts\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.792374 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/13bcfd35-b654-40e4-88d8-fc290ba96b14-hm-ports\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.792505 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/13bcfd35-b654-40e4-88d8-fc290ba96b14-config-data-merged\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.792640 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13bcfd35-b654-40e4-88d8-fc290ba96b14-config-data\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.793171 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/13bcfd35-b654-40e4-88d8-fc290ba96b14-config-data-merged\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.793573 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/13bcfd35-b654-40e4-88d8-fc290ba96b14-hm-ports\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.798997 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13bcfd35-b654-40e4-88d8-fc290ba96b14-config-data\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.799267 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13bcfd35-b654-40e4-88d8-fc290ba96b14-scripts\") pod \"octavia-rsyslog-ckzbv\" (UID: \"13bcfd35-b654-40e4-88d8-fc290ba96b14\") " pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:53 crc kubenswrapper[4871]: I1007 23:48:53.944082 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.502004 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-678599687f-m9kc4"] Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.504671 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.510519 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.510609 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-m9kc4"] Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.561727 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-ckzbv"] Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.608244 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/e12e730e-5fa7-4740-ac1d-ee08c0d54579-amphora-image\") pod \"octavia-image-upload-678599687f-m9kc4\" (UID: \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\") " pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.608693 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e12e730e-5fa7-4740-ac1d-ee08c0d54579-httpd-config\") pod \"octavia-image-upload-678599687f-m9kc4\" (UID: \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\") " pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.711373 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e12e730e-5fa7-4740-ac1d-ee08c0d54579-httpd-config\") pod \"octavia-image-upload-678599687f-m9kc4\" (UID: \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\") " pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.711511 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/e12e730e-5fa7-4740-ac1d-ee08c0d54579-amphora-image\") pod \"octavia-image-upload-678599687f-m9kc4\" (UID: \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\") " pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.712220 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/e12e730e-5fa7-4740-ac1d-ee08c0d54579-amphora-image\") pod \"octavia-image-upload-678599687f-m9kc4\" (UID: \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\") " pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.718991 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e12e730e-5fa7-4740-ac1d-ee08c0d54579-httpd-config\") pod \"octavia-image-upload-678599687f-m9kc4\" (UID: \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\") " pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:48:54 crc kubenswrapper[4871]: I1007 23:48:54.834162 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:48:55 crc kubenswrapper[4871]: I1007 23:48:55.258861 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-m9kc4"] Oct 07 23:48:55 crc kubenswrapper[4871]: W1007 23:48:55.264422 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode12e730e_5fa7_4740_ac1d_ee08c0d54579.slice/crio-9ccdc431e9cf4bb7c2ac89d195d86c5ced0bcd9f610df2dab8e891d39053f4f7 WatchSource:0}: Error finding container 9ccdc431e9cf4bb7c2ac89d195d86c5ced0bcd9f610df2dab8e891d39053f4f7: Status 404 returned error can't find the container with id 9ccdc431e9cf4bb7c2ac89d195d86c5ced0bcd9f610df2dab8e891d39053f4f7 Oct 07 23:48:55 crc kubenswrapper[4871]: I1007 23:48:55.496660 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-m9kc4" event={"ID":"e12e730e-5fa7-4740-ac1d-ee08c0d54579","Type":"ContainerStarted","Data":"9ccdc431e9cf4bb7c2ac89d195d86c5ced0bcd9f610df2dab8e891d39053f4f7"} Oct 07 23:48:55 crc kubenswrapper[4871]: I1007 23:48:55.499037 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-ckzbv" event={"ID":"13bcfd35-b654-40e4-88d8-fc290ba96b14","Type":"ContainerStarted","Data":"5e81407d086f79e855ef5dced39b103c127c0f22053f0d17076d4d067a06bae6"} Oct 07 23:48:56 crc kubenswrapper[4871]: I1007 23:48:56.511062 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-ckzbv" event={"ID":"13bcfd35-b654-40e4-88d8-fc290ba96b14","Type":"ContainerStarted","Data":"8ad6f9fcb25917c971d50bf39b17df8b82da725a8a458fd620cc0dd7e8c0ca57"} Oct 07 23:48:58 crc kubenswrapper[4871]: I1007 23:48:58.530065 4871 generic.go:334] "Generic (PLEG): container finished" podID="13bcfd35-b654-40e4-88d8-fc290ba96b14" containerID="8ad6f9fcb25917c971d50bf39b17df8b82da725a8a458fd620cc0dd7e8c0ca57" exitCode=0 Oct 07 23:48:58 crc kubenswrapper[4871]: I1007 23:48:58.530170 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-ckzbv" event={"ID":"13bcfd35-b654-40e4-88d8-fc290ba96b14","Type":"ContainerDied","Data":"8ad6f9fcb25917c971d50bf39b17df8b82da725a8a458fd620cc0dd7e8c0ca57"} Oct 07 23:48:58 crc kubenswrapper[4871]: I1007 23:48:58.533398 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 23:49:00 crc kubenswrapper[4871]: I1007 23:49:00.832500 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-plj8s"] Oct 07 23:49:00 crc kubenswrapper[4871]: I1007 23:49:00.834760 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:00 crc kubenswrapper[4871]: I1007 23:49:00.842042 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Oct 07 23:49:00 crc kubenswrapper[4871]: I1007 23:49:00.847388 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-plj8s"] Oct 07 23:49:00 crc kubenswrapper[4871]: I1007 23:49:00.939012 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-config-data\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:00 crc kubenswrapper[4871]: I1007 23:49:00.939056 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-scripts\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:00 crc kubenswrapper[4871]: I1007 23:49:00.939080 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/44085888-b090-4d53-964a-def7da4c1b00-config-data-merged\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:00 crc kubenswrapper[4871]: I1007 23:49:00.939129 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-combined-ca-bundle\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:01 crc kubenswrapper[4871]: I1007 23:49:01.042148 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-config-data\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:01 crc kubenswrapper[4871]: I1007 23:49:01.042577 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-scripts\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:01 crc kubenswrapper[4871]: I1007 23:49:01.042619 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/44085888-b090-4d53-964a-def7da4c1b00-config-data-merged\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:01 crc kubenswrapper[4871]: I1007 23:49:01.042654 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-combined-ca-bundle\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:01 crc kubenswrapper[4871]: I1007 23:49:01.043350 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/44085888-b090-4d53-964a-def7da4c1b00-config-data-merged\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:01 crc kubenswrapper[4871]: I1007 23:49:01.048226 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-config-data\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:01 crc kubenswrapper[4871]: I1007 23:49:01.048574 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-scripts\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:01 crc kubenswrapper[4871]: I1007 23:49:01.064909 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-combined-ca-bundle\") pod \"octavia-db-sync-plj8s\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:01 crc kubenswrapper[4871]: I1007 23:49:01.161735 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:04 crc kubenswrapper[4871]: W1007 23:49:04.791905 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44085888_b090_4d53_964a_def7da4c1b00.slice/crio-34b5548a11fe3b65c3f6882a03915e819ff9a386fb565d177f32d73ed8f1f8ff WatchSource:0}: Error finding container 34b5548a11fe3b65c3f6882a03915e819ff9a386fb565d177f32d73ed8f1f8ff: Status 404 returned error can't find the container with id 34b5548a11fe3b65c3f6882a03915e819ff9a386fb565d177f32d73ed8f1f8ff Oct 07 23:49:04 crc kubenswrapper[4871]: I1007 23:49:04.795261 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-plj8s"] Oct 07 23:49:05 crc kubenswrapper[4871]: I1007 23:49:05.611204 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-plj8s" event={"ID":"44085888-b090-4d53-964a-def7da4c1b00","Type":"ContainerStarted","Data":"34b5548a11fe3b65c3f6882a03915e819ff9a386fb565d177f32d73ed8f1f8ff"} Oct 07 23:49:05 crc kubenswrapper[4871]: I1007 23:49:05.615122 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-ckzbv" event={"ID":"13bcfd35-b654-40e4-88d8-fc290ba96b14","Type":"ContainerStarted","Data":"94e5a26592ae7effda4e2770813273b3bf249d9495d1ffc47769d60771bd3689"} Oct 07 23:49:05 crc kubenswrapper[4871]: I1007 23:49:05.616316 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:49:05 crc kubenswrapper[4871]: I1007 23:49:05.617721 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-m9kc4" event={"ID":"e12e730e-5fa7-4740-ac1d-ee08c0d54579","Type":"ContainerStarted","Data":"ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c"} Oct 07 23:49:05 crc kubenswrapper[4871]: I1007 23:49:05.653604 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-ckzbv" podStartSLOduration=2.865964107 podStartE2EDuration="12.653550272s" podCreationTimestamp="2025-10-07 23:48:53 +0000 UTC" firstStartedPulling="2025-10-07 23:48:54.544494791 +0000 UTC m=+6008.347192864" lastFinishedPulling="2025-10-07 23:49:04.332080916 +0000 UTC m=+6018.134779029" observedRunningTime="2025-10-07 23:49:05.641977316 +0000 UTC m=+6019.444675389" watchObservedRunningTime="2025-10-07 23:49:05.653550272 +0000 UTC m=+6019.456248385" Oct 07 23:49:06 crc kubenswrapper[4871]: I1007 23:49:06.634228 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-plj8s" event={"ID":"44085888-b090-4d53-964a-def7da4c1b00","Type":"ContainerStarted","Data":"d9085652688471218090da443ee3015394e601b9b013e425fcfe2d9ff96a1e22"} Oct 07 23:49:06 crc kubenswrapper[4871]: I1007 23:49:06.770015 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:49:06 crc kubenswrapper[4871]: I1007 23:49:06.945525 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-7bb5c546c4-ms9lr" Oct 07 23:49:07 crc kubenswrapper[4871]: I1007 23:49:07.645721 4871 generic.go:334] "Generic (PLEG): container finished" podID="44085888-b090-4d53-964a-def7da4c1b00" containerID="d9085652688471218090da443ee3015394e601b9b013e425fcfe2d9ff96a1e22" exitCode=0 Oct 07 23:49:07 crc kubenswrapper[4871]: I1007 23:49:07.645855 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-plj8s" event={"ID":"44085888-b090-4d53-964a-def7da4c1b00","Type":"ContainerDied","Data":"d9085652688471218090da443ee3015394e601b9b013e425fcfe2d9ff96a1e22"} Oct 07 23:49:07 crc kubenswrapper[4871]: I1007 23:49:07.648091 4871 generic.go:334] "Generic (PLEG): container finished" podID="e12e730e-5fa7-4740-ac1d-ee08c0d54579" containerID="ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c" exitCode=0 Oct 07 23:49:07 crc kubenswrapper[4871]: I1007 23:49:07.649664 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-m9kc4" event={"ID":"e12e730e-5fa7-4740-ac1d-ee08c0d54579","Type":"ContainerDied","Data":"ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c"} Oct 07 23:49:08 crc kubenswrapper[4871]: I1007 23:49:08.662481 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-m9kc4" event={"ID":"e12e730e-5fa7-4740-ac1d-ee08c0d54579","Type":"ContainerStarted","Data":"db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f"} Oct 07 23:49:09 crc kubenswrapper[4871]: I1007 23:49:09.708379 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-678599687f-m9kc4" podStartSLOduration=6.561055064 podStartE2EDuration="15.708360456s" podCreationTimestamp="2025-10-07 23:48:54 +0000 UTC" firstStartedPulling="2025-10-07 23:48:55.267053626 +0000 UTC m=+6009.069751699" lastFinishedPulling="2025-10-07 23:49:04.414358988 +0000 UTC m=+6018.217057091" observedRunningTime="2025-10-07 23:49:09.703319493 +0000 UTC m=+6023.506017606" watchObservedRunningTime="2025-10-07 23:49:09.708360456 +0000 UTC m=+6023.511058539" Oct 07 23:49:10 crc kubenswrapper[4871]: I1007 23:49:10.686287 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-plj8s" event={"ID":"44085888-b090-4d53-964a-def7da4c1b00","Type":"ContainerStarted","Data":"cae258903f12b66d45ac5b69b7b9a291017fce79fdb679330c3e6c17a1a2e33e"} Oct 07 23:49:10 crc kubenswrapper[4871]: I1007 23:49:10.721129 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-plj8s" podStartSLOduration=10.721111952 podStartE2EDuration="10.721111952s" podCreationTimestamp="2025-10-07 23:49:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:49:10.715267868 +0000 UTC m=+6024.517965941" watchObservedRunningTime="2025-10-07 23:49:10.721111952 +0000 UTC m=+6024.523810025" Oct 07 23:49:13 crc kubenswrapper[4871]: I1007 23:49:13.743966 4871 generic.go:334] "Generic (PLEG): container finished" podID="44085888-b090-4d53-964a-def7da4c1b00" containerID="cae258903f12b66d45ac5b69b7b9a291017fce79fdb679330c3e6c17a1a2e33e" exitCode=0 Oct 07 23:49:13 crc kubenswrapper[4871]: I1007 23:49:13.744827 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-plj8s" event={"ID":"44085888-b090-4d53-964a-def7da4c1b00","Type":"ContainerDied","Data":"cae258903f12b66d45ac5b69b7b9a291017fce79fdb679330c3e6c17a1a2e33e"} Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.119895 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.263756 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-config-data\") pod \"44085888-b090-4d53-964a-def7da4c1b00\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.264123 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-combined-ca-bundle\") pod \"44085888-b090-4d53-964a-def7da4c1b00\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.264183 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/44085888-b090-4d53-964a-def7da4c1b00-config-data-merged\") pod \"44085888-b090-4d53-964a-def7da4c1b00\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.264239 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-scripts\") pod \"44085888-b090-4d53-964a-def7da4c1b00\" (UID: \"44085888-b090-4d53-964a-def7da4c1b00\") " Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.270526 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-config-data" (OuterVolumeSpecName: "config-data") pod "44085888-b090-4d53-964a-def7da4c1b00" (UID: "44085888-b090-4d53-964a-def7da4c1b00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.271116 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-scripts" (OuterVolumeSpecName: "scripts") pod "44085888-b090-4d53-964a-def7da4c1b00" (UID: "44085888-b090-4d53-964a-def7da4c1b00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.298555 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44085888-b090-4d53-964a-def7da4c1b00" (UID: "44085888-b090-4d53-964a-def7da4c1b00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.310224 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44085888-b090-4d53-964a-def7da4c1b00-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "44085888-b090-4d53-964a-def7da4c1b00" (UID: "44085888-b090-4d53-964a-def7da4c1b00"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.366443 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.366813 4871 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/44085888-b090-4d53-964a-def7da4c1b00-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.366828 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.366839 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44085888-b090-4d53-964a-def7da4c1b00-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.765961 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-plj8s" event={"ID":"44085888-b090-4d53-964a-def7da4c1b00","Type":"ContainerDied","Data":"34b5548a11fe3b65c3f6882a03915e819ff9a386fb565d177f32d73ed8f1f8ff"} Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.766027 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-plj8s" Oct 07 23:49:15 crc kubenswrapper[4871]: I1007 23:49:15.766037 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34b5548a11fe3b65c3f6882a03915e819ff9a386fb565d177f32d73ed8f1f8ff" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.552465 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7k88q"] Oct 07 23:49:21 crc kubenswrapper[4871]: E1007 23:49:21.554013 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44085888-b090-4d53-964a-def7da4c1b00" containerName="octavia-db-sync" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.554026 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="44085888-b090-4d53-964a-def7da4c1b00" containerName="octavia-db-sync" Oct 07 23:49:21 crc kubenswrapper[4871]: E1007 23:49:21.554068 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44085888-b090-4d53-964a-def7da4c1b00" containerName="init" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.554075 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="44085888-b090-4d53-964a-def7da4c1b00" containerName="init" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.554482 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="44085888-b090-4d53-964a-def7da4c1b00" containerName="octavia-db-sync" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.557495 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.566571 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7k88q"] Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.714499 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-utilities\") pod \"redhat-operators-7k88q\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.714660 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpjb5\" (UniqueName: \"kubernetes.io/projected/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-kube-api-access-kpjb5\") pod \"redhat-operators-7k88q\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.714747 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-catalog-content\") pod \"redhat-operators-7k88q\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.816212 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpjb5\" (UniqueName: \"kubernetes.io/projected/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-kube-api-access-kpjb5\") pod \"redhat-operators-7k88q\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.816523 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-catalog-content\") pod \"redhat-operators-7k88q\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.816649 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-utilities\") pod \"redhat-operators-7k88q\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.817297 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-utilities\") pod \"redhat-operators-7k88q\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.817437 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-catalog-content\") pod \"redhat-operators-7k88q\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.839581 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpjb5\" (UniqueName: \"kubernetes.io/projected/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-kube-api-access-kpjb5\") pod \"redhat-operators-7k88q\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:21 crc kubenswrapper[4871]: I1007 23:49:21.921816 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:22 crc kubenswrapper[4871]: I1007 23:49:22.386370 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7k88q"] Oct 07 23:49:22 crc kubenswrapper[4871]: I1007 23:49:22.843866 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7k88q" event={"ID":"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37","Type":"ContainerStarted","Data":"ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778"} Oct 07 23:49:22 crc kubenswrapper[4871]: I1007 23:49:22.844252 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7k88q" event={"ID":"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37","Type":"ContainerStarted","Data":"8bb8d95b19b8e0299d7993c2dcf3ea51d02cfe24c58eec7f5e0ee96343caa74b"} Oct 07 23:49:23 crc kubenswrapper[4871]: I1007 23:49:23.864151 4871 generic.go:334] "Generic (PLEG): container finished" podID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerID="ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778" exitCode=0 Oct 07 23:49:23 crc kubenswrapper[4871]: I1007 23:49:23.864223 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7k88q" event={"ID":"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37","Type":"ContainerDied","Data":"ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778"} Oct 07 23:49:23 crc kubenswrapper[4871]: I1007 23:49:23.978728 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-ckzbv" Oct 07 23:49:25 crc kubenswrapper[4871]: I1007 23:49:25.883012 4871 generic.go:334] "Generic (PLEG): container finished" podID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerID="dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44" exitCode=0 Oct 07 23:49:25 crc kubenswrapper[4871]: I1007 23:49:25.883095 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7k88q" event={"ID":"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37","Type":"ContainerDied","Data":"dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44"} Oct 07 23:49:27 crc kubenswrapper[4871]: I1007 23:49:27.038326 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-6jsb5"] Oct 07 23:49:27 crc kubenswrapper[4871]: I1007 23:49:27.048486 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-6jsb5"] Oct 07 23:49:28 crc kubenswrapper[4871]: I1007 23:49:28.925945 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7k88q" event={"ID":"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37","Type":"ContainerStarted","Data":"5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394"} Oct 07 23:49:28 crc kubenswrapper[4871]: I1007 23:49:28.959169 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7k88q" podStartSLOduration=3.863066177 podStartE2EDuration="7.95914621s" podCreationTimestamp="2025-10-07 23:49:21 +0000 UTC" firstStartedPulling="2025-10-07 23:49:23.866476669 +0000 UTC m=+6037.669174742" lastFinishedPulling="2025-10-07 23:49:27.962556692 +0000 UTC m=+6041.765254775" observedRunningTime="2025-10-07 23:49:28.945271484 +0000 UTC m=+6042.747969597" watchObservedRunningTime="2025-10-07 23:49:28.95914621 +0000 UTC m=+6042.761844283" Oct 07 23:49:29 crc kubenswrapper[4871]: I1007 23:49:29.002601 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80322d0e-4017-4540-8ec5-98fa13ed2917" path="/var/lib/kubelet/pods/80322d0e-4017-4540-8ec5-98fa13ed2917/volumes" Oct 07 23:49:31 crc kubenswrapper[4871]: I1007 23:49:31.923800 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:31 crc kubenswrapper[4871]: I1007 23:49:31.924606 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:49:32 crc kubenswrapper[4871]: I1007 23:49:32.996070 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7k88q" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="registry-server" probeResult="failure" output=< Oct 07 23:49:32 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 23:49:32 crc kubenswrapper[4871]: > Oct 07 23:49:37 crc kubenswrapper[4871]: I1007 23:49:37.039477 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-0c02-account-create-qhdtz"] Oct 07 23:49:37 crc kubenswrapper[4871]: I1007 23:49:37.049832 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-0c02-account-create-qhdtz"] Oct 07 23:49:38 crc kubenswrapper[4871]: I1007 23:49:38.996521 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d" path="/var/lib/kubelet/pods/c7f1fe2a-9790-4c50-82c0-f2b8225d2d0d/volumes" Oct 07 23:49:42 crc kubenswrapper[4871]: I1007 23:49:42.973022 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7k88q" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="registry-server" probeResult="failure" output=< Oct 07 23:49:42 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 23:49:42 crc kubenswrapper[4871]: > Oct 07 23:49:44 crc kubenswrapper[4871]: I1007 23:49:44.035508 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dkp8g"] Oct 07 23:49:44 crc kubenswrapper[4871]: I1007 23:49:44.056875 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dkp8g"] Oct 07 23:49:44 crc kubenswrapper[4871]: I1007 23:49:44.472700 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-678599687f-m9kc4"] Oct 07 23:49:44 crc kubenswrapper[4871]: I1007 23:49:44.472961 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-678599687f-m9kc4" podUID="e12e730e-5fa7-4740-ac1d-ee08c0d54579" containerName="octavia-amphora-httpd" containerID="cri-o://db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f" gracePeriod=30 Oct 07 23:49:44 crc kubenswrapper[4871]: I1007 23:49:44.960318 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:49:44 crc kubenswrapper[4871]: I1007 23:49:44.993565 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="839e7e97-bf4c-4436-9fa5-9c45426422ff" path="/var/lib/kubelet/pods/839e7e97-bf4c-4436-9fa5-9c45426422ff/volumes" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.092214 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/e12e730e-5fa7-4740-ac1d-ee08c0d54579-amphora-image\") pod \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\" (UID: \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\") " Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.098952 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e12e730e-5fa7-4740-ac1d-ee08c0d54579-httpd-config\") pod \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\" (UID: \"e12e730e-5fa7-4740-ac1d-ee08c0d54579\") " Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.130628 4871 generic.go:334] "Generic (PLEG): container finished" podID="e12e730e-5fa7-4740-ac1d-ee08c0d54579" containerID="db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f" exitCode=0 Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.130876 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-m9kc4" event={"ID":"e12e730e-5fa7-4740-ac1d-ee08c0d54579","Type":"ContainerDied","Data":"db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f"} Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.131252 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-m9kc4" event={"ID":"e12e730e-5fa7-4740-ac1d-ee08c0d54579","Type":"ContainerDied","Data":"9ccdc431e9cf4bb7c2ac89d195d86c5ced0bcd9f610df2dab8e891d39053f4f7"} Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.131309 4871 scope.go:117] "RemoveContainer" containerID="db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.131339 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-m9kc4" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.131610 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e12e730e-5fa7-4740-ac1d-ee08c0d54579-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e12e730e-5fa7-4740-ac1d-ee08c0d54579" (UID: "e12e730e-5fa7-4740-ac1d-ee08c0d54579"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.142778 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e12e730e-5fa7-4740-ac1d-ee08c0d54579-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "e12e730e-5fa7-4740-ac1d-ee08c0d54579" (UID: "e12e730e-5fa7-4740-ac1d-ee08c0d54579"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.201830 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e12e730e-5fa7-4740-ac1d-ee08c0d54579-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.202166 4871 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/e12e730e-5fa7-4740-ac1d-ee08c0d54579-amphora-image\") on node \"crc\" DevicePath \"\"" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.223066 4871 scope.go:117] "RemoveContainer" containerID="ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.245372 4871 scope.go:117] "RemoveContainer" containerID="db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f" Oct 07 23:49:45 crc kubenswrapper[4871]: E1007 23:49:45.247737 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f\": container with ID starting with db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f not found: ID does not exist" containerID="db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.247771 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f"} err="failed to get container status \"db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f\": rpc error: code = NotFound desc = could not find container \"db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f\": container with ID starting with db5d36c5d8cc514510742d598c240e1ad6a40245e31a30547f77c927e00d9f8f not found: ID does not exist" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.247815 4871 scope.go:117] "RemoveContainer" containerID="ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c" Oct 07 23:49:45 crc kubenswrapper[4871]: E1007 23:49:45.248194 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c\": container with ID starting with ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c not found: ID does not exist" containerID="ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.248241 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c"} err="failed to get container status \"ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c\": rpc error: code = NotFound desc = could not find container \"ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c\": container with ID starting with ccd606f946e61dab355c6a6c65b5a8347ec03e0b3c21ba1ea08a2f4aaebf075c not found: ID does not exist" Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.461048 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-678599687f-m9kc4"] Oct 07 23:49:45 crc kubenswrapper[4871]: I1007 23:49:45.470298 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-678599687f-m9kc4"] Oct 07 23:49:47 crc kubenswrapper[4871]: I1007 23:49:47.005566 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e12e730e-5fa7-4740-ac1d-ee08c0d54579" path="/var/lib/kubelet/pods/e12e730e-5fa7-4740-ac1d-ee08c0d54579/volumes" Oct 07 23:49:52 crc kubenswrapper[4871]: I1007 23:49:52.991567 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7k88q" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="registry-server" probeResult="failure" output=< Oct 07 23:49:52 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 23:49:52 crc kubenswrapper[4871]: > Oct 07 23:49:58 crc kubenswrapper[4871]: I1007 23:49:58.187684 4871 scope.go:117] "RemoveContainer" containerID="410417bc7820ae652758e1d45f46751f7247c7852fb25ca8ddd575ca2c2baf2d" Oct 07 23:49:58 crc kubenswrapper[4871]: I1007 23:49:58.246403 4871 scope.go:117] "RemoveContainer" containerID="85a8da9f1c3210181112ade3024b0a769ed0e283f5389f18d35e8f5b80b0727f" Oct 07 23:49:58 crc kubenswrapper[4871]: I1007 23:49:58.323729 4871 scope.go:117] "RemoveContainer" containerID="5af235543317d416b3ea0b9fb8edceb8e1d2aca6a85e1c4f511e1521b38236f9" Oct 07 23:49:58 crc kubenswrapper[4871]: I1007 23:49:58.414485 4871 scope.go:117] "RemoveContainer" containerID="20765c19ad6595246dd2f00dadbd6ec9d43e91afd3aed34e564a56681fda5d61" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.798452 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-knthn"] Oct 07 23:50:00 crc kubenswrapper[4871]: E1007 23:50:00.802264 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12e730e-5fa7-4740-ac1d-ee08c0d54579" containerName="init" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.802357 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12e730e-5fa7-4740-ac1d-ee08c0d54579" containerName="init" Oct 07 23:50:00 crc kubenswrapper[4871]: E1007 23:50:00.802461 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e12e730e-5fa7-4740-ac1d-ee08c0d54579" containerName="octavia-amphora-httpd" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.802519 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e12e730e-5fa7-4740-ac1d-ee08c0d54579" containerName="octavia-amphora-httpd" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.802989 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e12e730e-5fa7-4740-ac1d-ee08c0d54579" containerName="octavia-amphora-httpd" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.805654 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.812020 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.828290 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.828458 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.836402 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-knthn"] Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.963939 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-combined-ca-bundle\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.964332 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-config-data\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.964430 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d730e692-7f26-4844-a01d-f79e38ced5fc-hm-ports\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.964498 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-amphora-certs\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.964586 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-scripts\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:00 crc kubenswrapper[4871]: I1007 23:50:00.964732 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d730e692-7f26-4844-a01d-f79e38ced5fc-config-data-merged\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.066880 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-combined-ca-bundle\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.066964 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-config-data\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.067030 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d730e692-7f26-4844-a01d-f79e38ced5fc-hm-ports\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.067065 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-amphora-certs\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.067107 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-scripts\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.067187 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d730e692-7f26-4844-a01d-f79e38ced5fc-config-data-merged\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.067946 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d730e692-7f26-4844-a01d-f79e38ced5fc-config-data-merged\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.068448 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d730e692-7f26-4844-a01d-f79e38ced5fc-hm-ports\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.073959 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-config-data\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.074598 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-amphora-certs\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.074903 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-scripts\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.088624 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d730e692-7f26-4844-a01d-f79e38ced5fc-combined-ca-bundle\") pod \"octavia-healthmanager-knthn\" (UID: \"d730e692-7f26-4844-a01d-f79e38ced5fc\") " pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.131819 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.719483 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-knthn"] Oct 07 23:50:01 crc kubenswrapper[4871]: I1007 23:50:01.986612 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:50:02 crc kubenswrapper[4871]: I1007 23:50:02.036744 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:50:02 crc kubenswrapper[4871]: I1007 23:50:02.222725 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7k88q"] Oct 07 23:50:02 crc kubenswrapper[4871]: I1007 23:50:02.316830 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-knthn" event={"ID":"d730e692-7f26-4844-a01d-f79e38ced5fc","Type":"ContainerStarted","Data":"bc0ca1654e4209a2cfb66d1e19d97c6085033e8e51d424a0508b6c46bb06d988"} Oct 07 23:50:02 crc kubenswrapper[4871]: I1007 23:50:02.316876 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-knthn" event={"ID":"d730e692-7f26-4844-a01d-f79e38ced5fc","Type":"ContainerStarted","Data":"76f957dac0001c2b4d3e36d7dfe517b956ce557195a11616d9c2cd27de1ee994"} Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.179889 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-f2v96"] Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.181722 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.183236 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.184349 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.194293 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-f2v96"] Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.310864 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-scripts\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.310929 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-amphora-certs\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.310959 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f01a2153-7671-41bf-9c79-7bf208956897-hm-ports\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.311030 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-config-data\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.311084 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f01a2153-7671-41bf-9c79-7bf208956897-config-data-merged\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.311130 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-combined-ca-bundle\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.324259 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7k88q" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="registry-server" containerID="cri-o://5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394" gracePeriod=2 Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.412719 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-combined-ca-bundle\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.412774 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-scripts\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.412835 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-amphora-certs\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.412871 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f01a2153-7671-41bf-9c79-7bf208956897-hm-ports\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.412943 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-config-data\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.413006 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f01a2153-7671-41bf-9c79-7bf208956897-config-data-merged\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.413496 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f01a2153-7671-41bf-9c79-7bf208956897-config-data-merged\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.415053 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f01a2153-7671-41bf-9c79-7bf208956897-hm-ports\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.419027 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-combined-ca-bundle\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.419646 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-scripts\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.420084 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-amphora-certs\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.420579 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01a2153-7671-41bf-9c79-7bf208956897-config-data\") pod \"octavia-housekeeping-f2v96\" (UID: \"f01a2153-7671-41bf-9c79-7bf208956897\") " pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.496673 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.829284 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.923545 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-catalog-content\") pod \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.923610 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpjb5\" (UniqueName: \"kubernetes.io/projected/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-kube-api-access-kpjb5\") pod \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.923773 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-utilities\") pod \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.924956 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-utilities" (OuterVolumeSpecName: "utilities") pod "b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" (UID: "b7b2b22e-1f30-41dd-ab97-bb8ad6318b37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:50:03 crc kubenswrapper[4871]: I1007 23:50:03.929413 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-kube-api-access-kpjb5" (OuterVolumeSpecName: "kube-api-access-kpjb5") pod "b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" (UID: "b7b2b22e-1f30-41dd-ab97-bb8ad6318b37"). InnerVolumeSpecName "kube-api-access-kpjb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.024669 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" (UID: "b7b2b22e-1f30-41dd-ab97-bb8ad6318b37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.025543 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-catalog-content\") pod \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\" (UID: \"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37\") " Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.026166 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpjb5\" (UniqueName: \"kubernetes.io/projected/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-kube-api-access-kpjb5\") on node \"crc\" DevicePath \"\"" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.026189 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:50:04 crc kubenswrapper[4871]: W1007 23:50:04.026273 4871 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37/volumes/kubernetes.io~empty-dir/catalog-content Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.026288 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" (UID: "b7b2b22e-1f30-41dd-ab97-bb8ad6318b37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.098866 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-f2v96"] Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.128590 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.348099 4871 generic.go:334] "Generic (PLEG): container finished" podID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerID="5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394" exitCode=0 Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.348205 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7k88q" event={"ID":"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37","Type":"ContainerDied","Data":"5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394"} Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.348250 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7k88q" event={"ID":"b7b2b22e-1f30-41dd-ab97-bb8ad6318b37","Type":"ContainerDied","Data":"8bb8d95b19b8e0299d7993c2dcf3ea51d02cfe24c58eec7f5e0ee96343caa74b"} Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.348277 4871 scope.go:117] "RemoveContainer" containerID="5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.348520 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7k88q" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.351975 4871 generic.go:334] "Generic (PLEG): container finished" podID="d730e692-7f26-4844-a01d-f79e38ced5fc" containerID="bc0ca1654e4209a2cfb66d1e19d97c6085033e8e51d424a0508b6c46bb06d988" exitCode=0 Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.352093 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-knthn" event={"ID":"d730e692-7f26-4844-a01d-f79e38ced5fc","Type":"ContainerDied","Data":"bc0ca1654e4209a2cfb66d1e19d97c6085033e8e51d424a0508b6c46bb06d988"} Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.362391 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-f2v96" event={"ID":"f01a2153-7671-41bf-9c79-7bf208956897","Type":"ContainerStarted","Data":"4f4de0a21043952ceef53338af061f10e997b801614aa0f7a44d41ae5bc6ba9a"} Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.376002 4871 scope.go:117] "RemoveContainer" containerID="dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.422947 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7k88q"] Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.439024 4871 scope.go:117] "RemoveContainer" containerID="ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.443501 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7k88q"] Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.546444 4871 scope.go:117] "RemoveContainer" containerID="5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394" Oct 07 23:50:04 crc kubenswrapper[4871]: E1007 23:50:04.547916 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394\": container with ID starting with 5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394 not found: ID does not exist" containerID="5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.547959 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394"} err="failed to get container status \"5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394\": rpc error: code = NotFound desc = could not find container \"5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394\": container with ID starting with 5201257cfcfdab76dd2052c1a3c3f4bfbaf225633317137cadea89b0a6e94394 not found: ID does not exist" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.547984 4871 scope.go:117] "RemoveContainer" containerID="dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44" Oct 07 23:50:04 crc kubenswrapper[4871]: E1007 23:50:04.548613 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44\": container with ID starting with dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44 not found: ID does not exist" containerID="dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.548639 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44"} err="failed to get container status \"dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44\": rpc error: code = NotFound desc = could not find container \"dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44\": container with ID starting with dc699176b3fa966b352699e9da844abb879559e0a3529968af9f5c69ff6ada44 not found: ID does not exist" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.548653 4871 scope.go:117] "RemoveContainer" containerID="ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778" Oct 07 23:50:04 crc kubenswrapper[4871]: E1007 23:50:04.549046 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778\": container with ID starting with ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778 not found: ID does not exist" containerID="ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.549066 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778"} err="failed to get container status \"ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778\": rpc error: code = NotFound desc = could not find container \"ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778\": container with ID starting with ea23341d49f6b0d35747bd3f0e564982c347f477204de985b08488b31411d778 not found: ID does not exist" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.781698 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-jnzlv"] Oct 07 23:50:04 crc kubenswrapper[4871]: E1007 23:50:04.782121 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="extract-utilities" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.782143 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="extract-utilities" Oct 07 23:50:04 crc kubenswrapper[4871]: E1007 23:50:04.782165 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="extract-content" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.782172 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="extract-content" Oct 07 23:50:04 crc kubenswrapper[4871]: E1007 23:50:04.782204 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="registry-server" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.782210 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="registry-server" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.782380 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" containerName="registry-server" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.784097 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.788317 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.788450 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.801062 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-jnzlv"] Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.949084 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-combined-ca-bundle\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.949134 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-scripts\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.949209 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8eeebe46-f544-4884-b33b-06ed31f5c6f7-config-data-merged\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.949241 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-config-data\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.949266 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-amphora-certs\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.949283 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8eeebe46-f544-4884-b33b-06ed31f5c6f7-hm-ports\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:04 crc kubenswrapper[4871]: I1007 23:50:04.994093 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7b2b22e-1f30-41dd-ab97-bb8ad6318b37" path="/var/lib/kubelet/pods/b7b2b22e-1f30-41dd-ab97-bb8ad6318b37/volumes" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.050591 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-amphora-certs\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.050634 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8eeebe46-f544-4884-b33b-06ed31f5c6f7-hm-ports\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.050734 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-combined-ca-bundle\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.050759 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-scripts\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.050865 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8eeebe46-f544-4884-b33b-06ed31f5c6f7-config-data-merged\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.050897 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-config-data\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.051466 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8eeebe46-f544-4884-b33b-06ed31f5c6f7-config-data-merged\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.052415 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8eeebe46-f544-4884-b33b-06ed31f5c6f7-hm-ports\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.067286 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-combined-ca-bundle\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.067913 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-amphora-certs\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.075781 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-config-data\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.082685 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eeebe46-f544-4884-b33b-06ed31f5c6f7-scripts\") pod \"octavia-worker-jnzlv\" (UID: \"8eeebe46-f544-4884-b33b-06ed31f5c6f7\") " pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.118501 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.375985 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-knthn" event={"ID":"d730e692-7f26-4844-a01d-f79e38ced5fc","Type":"ContainerStarted","Data":"01747ef35474db32a5bf4288ec77c9a47d88d3482ef465aae9cdf553208ff58b"} Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.377099 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:05 crc kubenswrapper[4871]: I1007 23:50:05.406919 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-knthn" podStartSLOduration=5.40689734 podStartE2EDuration="5.40689734s" podCreationTimestamp="2025-10-07 23:50:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:50:05.398335585 +0000 UTC m=+6079.201033658" watchObservedRunningTime="2025-10-07 23:50:05.40689734 +0000 UTC m=+6079.209595413" Oct 07 23:50:06 crc kubenswrapper[4871]: I1007 23:50:06.021398 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-jnzlv"] Oct 07 23:50:06 crc kubenswrapper[4871]: I1007 23:50:06.389320 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-jnzlv" event={"ID":"8eeebe46-f544-4884-b33b-06ed31f5c6f7","Type":"ContainerStarted","Data":"4ed35765ebf46eee369588e19a7e8c10fdf5ac3149e6a01346a989dd0698cf26"} Oct 07 23:50:06 crc kubenswrapper[4871]: I1007 23:50:06.393901 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-f2v96" event={"ID":"f01a2153-7671-41bf-9c79-7bf208956897","Type":"ContainerStarted","Data":"3534949b4862380a016a206f8b8e59be9dd9edcbe2b27ed7804611229c9a1098"} Oct 07 23:50:07 crc kubenswrapper[4871]: I1007 23:50:07.405232 4871 generic.go:334] "Generic (PLEG): container finished" podID="f01a2153-7671-41bf-9c79-7bf208956897" containerID="3534949b4862380a016a206f8b8e59be9dd9edcbe2b27ed7804611229c9a1098" exitCode=0 Oct 07 23:50:07 crc kubenswrapper[4871]: I1007 23:50:07.405507 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-f2v96" event={"ID":"f01a2153-7671-41bf-9c79-7bf208956897","Type":"ContainerDied","Data":"3534949b4862380a016a206f8b8e59be9dd9edcbe2b27ed7804611229c9a1098"} Oct 07 23:50:07 crc kubenswrapper[4871]: I1007 23:50:07.768894 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qg2jp"] Oct 07 23:50:07 crc kubenswrapper[4871]: I1007 23:50:07.771211 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:07 crc kubenswrapper[4871]: I1007 23:50:07.801131 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg2jp"] Oct 07 23:50:07 crc kubenswrapper[4871]: I1007 23:50:07.910199 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-catalog-content\") pod \"redhat-marketplace-qg2jp\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:07 crc kubenswrapper[4871]: I1007 23:50:07.910273 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-utilities\") pod \"redhat-marketplace-qg2jp\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:07 crc kubenswrapper[4871]: I1007 23:50:07.910469 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnc6c\" (UniqueName: \"kubernetes.io/projected/d56f53ee-f4ed-40fb-b940-518c5ef52c41-kube-api-access-pnc6c\") pod \"redhat-marketplace-qg2jp\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.012610 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnc6c\" (UniqueName: \"kubernetes.io/projected/d56f53ee-f4ed-40fb-b940-518c5ef52c41-kube-api-access-pnc6c\") pod \"redhat-marketplace-qg2jp\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.013118 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-catalog-content\") pod \"redhat-marketplace-qg2jp\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.013193 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-utilities\") pod \"redhat-marketplace-qg2jp\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.013821 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-catalog-content\") pod \"redhat-marketplace-qg2jp\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.013864 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-utilities\") pod \"redhat-marketplace-qg2jp\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.042265 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnc6c\" (UniqueName: \"kubernetes.io/projected/d56f53ee-f4ed-40fb-b940-518c5ef52c41-kube-api-access-pnc6c\") pod \"redhat-marketplace-qg2jp\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.104907 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.421196 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-jnzlv" event={"ID":"8eeebe46-f544-4884-b33b-06ed31f5c6f7","Type":"ContainerStarted","Data":"81c1392df0c5966c9e137c0ed34101bf33c0ca47b1838c2c28c70034895ea351"} Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.423760 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-f2v96" event={"ID":"f01a2153-7671-41bf-9c79-7bf208956897","Type":"ContainerStarted","Data":"68aad1a47f0d8c0bbfa7a3ed487a74160387e5bef043a74849d306c4a43afcab"} Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.423908 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.467008 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-f2v96" podStartSLOduration=4.061823148 podStartE2EDuration="5.466989113s" podCreationTimestamp="2025-10-07 23:50:03 +0000 UTC" firstStartedPulling="2025-10-07 23:50:04.111557375 +0000 UTC m=+6077.914255448" lastFinishedPulling="2025-10-07 23:50:05.51672332 +0000 UTC m=+6079.319421413" observedRunningTime="2025-10-07 23:50:08.458898011 +0000 UTC m=+6082.261596084" watchObservedRunningTime="2025-10-07 23:50:08.466989113 +0000 UTC m=+6082.269687186" Oct 07 23:50:08 crc kubenswrapper[4871]: I1007 23:50:08.562268 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg2jp"] Oct 07 23:50:08 crc kubenswrapper[4871]: W1007 23:50:08.567167 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd56f53ee_f4ed_40fb_b940_518c5ef52c41.slice/crio-6c46de6a010456ef083a395145ef72c355a36c861689a717ff90a16de2b51a24 WatchSource:0}: Error finding container 6c46de6a010456ef083a395145ef72c355a36c861689a717ff90a16de2b51a24: Status 404 returned error can't find the container with id 6c46de6a010456ef083a395145ef72c355a36c861689a717ff90a16de2b51a24 Oct 07 23:50:09 crc kubenswrapper[4871]: I1007 23:50:09.436340 4871 generic.go:334] "Generic (PLEG): container finished" podID="8eeebe46-f544-4884-b33b-06ed31f5c6f7" containerID="81c1392df0c5966c9e137c0ed34101bf33c0ca47b1838c2c28c70034895ea351" exitCode=0 Oct 07 23:50:09 crc kubenswrapper[4871]: I1007 23:50:09.436440 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-jnzlv" event={"ID":"8eeebe46-f544-4884-b33b-06ed31f5c6f7","Type":"ContainerDied","Data":"81c1392df0c5966c9e137c0ed34101bf33c0ca47b1838c2c28c70034895ea351"} Oct 07 23:50:09 crc kubenswrapper[4871]: I1007 23:50:09.440483 4871 generic.go:334] "Generic (PLEG): container finished" podID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerID="c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560" exitCode=0 Oct 07 23:50:09 crc kubenswrapper[4871]: I1007 23:50:09.440686 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg2jp" event={"ID":"d56f53ee-f4ed-40fb-b940-518c5ef52c41","Type":"ContainerDied","Data":"c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560"} Oct 07 23:50:09 crc kubenswrapper[4871]: I1007 23:50:09.440720 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg2jp" event={"ID":"d56f53ee-f4ed-40fb-b940-518c5ef52c41","Type":"ContainerStarted","Data":"6c46de6a010456ef083a395145ef72c355a36c861689a717ff90a16de2b51a24"} Oct 07 23:50:10 crc kubenswrapper[4871]: I1007 23:50:10.464185 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-jnzlv" event={"ID":"8eeebe46-f544-4884-b33b-06ed31f5c6f7","Type":"ContainerStarted","Data":"b3b2175997d3773b4c551572a7327b1089650dc0187f34d07836a1205a93721a"} Oct 07 23:50:10 crc kubenswrapper[4871]: I1007 23:50:10.464459 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:10 crc kubenswrapper[4871]: I1007 23:50:10.499379 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-jnzlv" podStartSLOduration=5.024141607 podStartE2EDuration="6.499358461s" podCreationTimestamp="2025-10-07 23:50:04 +0000 UTC" firstStartedPulling="2025-10-07 23:50:06.031661877 +0000 UTC m=+6079.834359950" lastFinishedPulling="2025-10-07 23:50:07.506878701 +0000 UTC m=+6081.309576804" observedRunningTime="2025-10-07 23:50:10.485671062 +0000 UTC m=+6084.288369145" watchObservedRunningTime="2025-10-07 23:50:10.499358461 +0000 UTC m=+6084.302056544" Oct 07 23:50:11 crc kubenswrapper[4871]: E1007 23:50:11.065279 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd56f53ee_f4ed_40fb_b940_518c5ef52c41.slice/crio-conmon-f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd56f53ee_f4ed_40fb_b940_518c5ef52c41.slice/crio-f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004.scope\": RecentStats: unable to find data in memory cache]" Oct 07 23:50:11 crc kubenswrapper[4871]: I1007 23:50:11.476249 4871 generic.go:334] "Generic (PLEG): container finished" podID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerID="f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004" exitCode=0 Oct 07 23:50:11 crc kubenswrapper[4871]: I1007 23:50:11.476312 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg2jp" event={"ID":"d56f53ee-f4ed-40fb-b940-518c5ef52c41","Type":"ContainerDied","Data":"f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004"} Oct 07 23:50:12 crc kubenswrapper[4871]: I1007 23:50:12.039140 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-xln4c"] Oct 07 23:50:12 crc kubenswrapper[4871]: I1007 23:50:12.054467 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-xln4c"] Oct 07 23:50:12 crc kubenswrapper[4871]: I1007 23:50:12.496211 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg2jp" event={"ID":"d56f53ee-f4ed-40fb-b940-518c5ef52c41","Type":"ContainerStarted","Data":"18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257"} Oct 07 23:50:12 crc kubenswrapper[4871]: I1007 23:50:12.518710 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qg2jp" podStartSLOduration=2.804987658 podStartE2EDuration="5.518691976s" podCreationTimestamp="2025-10-07 23:50:07 +0000 UTC" firstStartedPulling="2025-10-07 23:50:09.442842349 +0000 UTC m=+6083.245540432" lastFinishedPulling="2025-10-07 23:50:12.156546677 +0000 UTC m=+6085.959244750" observedRunningTime="2025-10-07 23:50:12.511166939 +0000 UTC m=+6086.313865012" watchObservedRunningTime="2025-10-07 23:50:12.518691976 +0000 UTC m=+6086.321390049" Oct 07 23:50:12 crc kubenswrapper[4871]: I1007 23:50:12.993348 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50de0223-7b4d-4035-818d-1d238a83dd89" path="/var/lib/kubelet/pods/50de0223-7b4d-4035-818d-1d238a83dd89/volumes" Oct 07 23:50:16 crc kubenswrapper[4871]: I1007 23:50:16.165314 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-knthn" Oct 07 23:50:18 crc kubenswrapper[4871]: I1007 23:50:18.105841 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:18 crc kubenswrapper[4871]: I1007 23:50:18.106420 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:18 crc kubenswrapper[4871]: I1007 23:50:18.160580 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:18 crc kubenswrapper[4871]: I1007 23:50:18.526986 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-f2v96" Oct 07 23:50:18 crc kubenswrapper[4871]: I1007 23:50:18.669654 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:18 crc kubenswrapper[4871]: I1007 23:50:18.745786 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg2jp"] Oct 07 23:50:20 crc kubenswrapper[4871]: I1007 23:50:20.174150 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-jnzlv" Oct 07 23:50:20 crc kubenswrapper[4871]: I1007 23:50:20.596965 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qg2jp" podUID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerName="registry-server" containerID="cri-o://18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257" gracePeriod=2 Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.121636 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.199266 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-catalog-content\") pod \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.199351 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnc6c\" (UniqueName: \"kubernetes.io/projected/d56f53ee-f4ed-40fb-b940-518c5ef52c41-kube-api-access-pnc6c\") pod \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.199566 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-utilities\") pod \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\" (UID: \"d56f53ee-f4ed-40fb-b940-518c5ef52c41\") " Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.200971 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-utilities" (OuterVolumeSpecName: "utilities") pod "d56f53ee-f4ed-40fb-b940-518c5ef52c41" (UID: "d56f53ee-f4ed-40fb-b940-518c5ef52c41"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.211834 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d56f53ee-f4ed-40fb-b940-518c5ef52c41" (UID: "d56f53ee-f4ed-40fb-b940-518c5ef52c41"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.215687 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d56f53ee-f4ed-40fb-b940-518c5ef52c41-kube-api-access-pnc6c" (OuterVolumeSpecName: "kube-api-access-pnc6c") pod "d56f53ee-f4ed-40fb-b940-518c5ef52c41" (UID: "d56f53ee-f4ed-40fb-b940-518c5ef52c41"). InnerVolumeSpecName "kube-api-access-pnc6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.302893 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.302933 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56f53ee-f4ed-40fb-b940-518c5ef52c41-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.302948 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnc6c\" (UniqueName: \"kubernetes.io/projected/d56f53ee-f4ed-40fb-b940-518c5ef52c41-kube-api-access-pnc6c\") on node \"crc\" DevicePath \"\"" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.609607 4871 generic.go:334] "Generic (PLEG): container finished" podID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerID="18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257" exitCode=0 Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.609667 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg2jp" event={"ID":"d56f53ee-f4ed-40fb-b940-518c5ef52c41","Type":"ContainerDied","Data":"18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257"} Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.609706 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qg2jp" event={"ID":"d56f53ee-f4ed-40fb-b940-518c5ef52c41","Type":"ContainerDied","Data":"6c46de6a010456ef083a395145ef72c355a36c861689a717ff90a16de2b51a24"} Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.609734 4871 scope.go:117] "RemoveContainer" containerID="18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.609928 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qg2jp" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.650048 4871 scope.go:117] "RemoveContainer" containerID="f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.668968 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg2jp"] Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.677476 4871 scope.go:117] "RemoveContainer" containerID="c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.679633 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qg2jp"] Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.730744 4871 scope.go:117] "RemoveContainer" containerID="18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257" Oct 07 23:50:21 crc kubenswrapper[4871]: E1007 23:50:21.731286 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257\": container with ID starting with 18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257 not found: ID does not exist" containerID="18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.731362 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257"} err="failed to get container status \"18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257\": rpc error: code = NotFound desc = could not find container \"18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257\": container with ID starting with 18f717c13977b39899b8d496eae1ba3d81680ea8dd66b814d2f0f4a324a57257 not found: ID does not exist" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.731393 4871 scope.go:117] "RemoveContainer" containerID="f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004" Oct 07 23:50:21 crc kubenswrapper[4871]: E1007 23:50:21.731912 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004\": container with ID starting with f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004 not found: ID does not exist" containerID="f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.732002 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004"} err="failed to get container status \"f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004\": rpc error: code = NotFound desc = could not find container \"f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004\": container with ID starting with f16cba03af3b5cb25184aed8ed2ed01a1fe4138a8529596aac5d03a9aa200004 not found: ID does not exist" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.732047 4871 scope.go:117] "RemoveContainer" containerID="c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560" Oct 07 23:50:21 crc kubenswrapper[4871]: E1007 23:50:21.732419 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560\": container with ID starting with c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560 not found: ID does not exist" containerID="c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560" Oct 07 23:50:21 crc kubenswrapper[4871]: I1007 23:50:21.732474 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560"} err="failed to get container status \"c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560\": rpc error: code = NotFound desc = could not find container \"c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560\": container with ID starting with c78a1a28248b4d6292add7098f5ed6b9dd38b51aca47a2cefc69900c6fd64560 not found: ID does not exist" Oct 07 23:50:22 crc kubenswrapper[4871]: I1007 23:50:22.033146 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2101-account-create-f8565"] Oct 07 23:50:22 crc kubenswrapper[4871]: I1007 23:50:22.044485 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2101-account-create-f8565"] Oct 07 23:50:22 crc kubenswrapper[4871]: I1007 23:50:22.997450 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99101738-90d6-4b8a-98b3-10e9af17c641" path="/var/lib/kubelet/pods/99101738-90d6-4b8a-98b3-10e9af17c641/volumes" Oct 07 23:50:22 crc kubenswrapper[4871]: I1007 23:50:22.998977 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" path="/var/lib/kubelet/pods/d56f53ee-f4ed-40fb-b940-518c5ef52c41/volumes" Oct 07 23:50:31 crc kubenswrapper[4871]: I1007 23:50:31.028399 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kf9sd"] Oct 07 23:50:31 crc kubenswrapper[4871]: I1007 23:50:31.041585 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kf9sd"] Oct 07 23:50:32 crc kubenswrapper[4871]: I1007 23:50:32.994306 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45faaa8f-9462-4793-90ab-b22c394e7177" path="/var/lib/kubelet/pods/45faaa8f-9462-4793-90ab-b22c394e7177/volumes" Oct 07 23:50:58 crc kubenswrapper[4871]: I1007 23:50:58.587372 4871 scope.go:117] "RemoveContainer" containerID="e020cbc38f62809ceda30ee28b1b9e3c053ec981ebb85de2d43ab9d9a309ffdf" Oct 07 23:50:58 crc kubenswrapper[4871]: I1007 23:50:58.634231 4871 scope.go:117] "RemoveContainer" containerID="6de56f1138da3a53c6e3bff9654a0471c09c87d61fd117357e77f6098e352504" Oct 07 23:50:58 crc kubenswrapper[4871]: I1007 23:50:58.687065 4871 scope.go:117] "RemoveContainer" containerID="49a9e8a177238115f373d908ceb2588e15819b9848026176315cf2c5df2eed61" Oct 07 23:50:58 crc kubenswrapper[4871]: I1007 23:50:58.758883 4871 scope.go:117] "RemoveContainer" containerID="daa5ddfe41d3c442593ac170da5ddcd02661e3fcfa7e169934d7c3ddfa038f42" Oct 07 23:50:58 crc kubenswrapper[4871]: I1007 23:50:58.785101 4871 scope.go:117] "RemoveContainer" containerID="eb074beb4c7ef7b670094238e47500474752585e6a2999234668260465f87d8b" Oct 07 23:51:05 crc kubenswrapper[4871]: I1007 23:51:05.512385 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:51:05 crc kubenswrapper[4871]: I1007 23:51:05.514090 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.794869 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d94b76c57-wjk4c"] Oct 07 23:51:11 crc kubenswrapper[4871]: E1007 23:51:11.795745 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerName="extract-content" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.795760 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerName="extract-content" Oct 07 23:51:11 crc kubenswrapper[4871]: E1007 23:51:11.795829 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerName="registry-server" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.795841 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerName="registry-server" Oct 07 23:51:11 crc kubenswrapper[4871]: E1007 23:51:11.795860 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerName="extract-utilities" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.795871 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerName="extract-utilities" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.796101 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d56f53ee-f4ed-40fb-b940-518c5ef52c41" containerName="registry-server" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.797416 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.804749 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.805424 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.805775 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.806203 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-v8ll9" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.835117 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d94b76c57-wjk4c"] Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.907599 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.908050 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerName="glance-log" containerID="cri-o://a19d16bbb1da455b8bd1a86f2aecf55b9b15a8bfd8d48b6c2852c587c419ac82" gracePeriod=30 Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.908499 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerName="glance-httpd" containerID="cri-o://d0811bd3822ca6ebeb120cdc6461892d29a96b4c5c061b4fc658953ddaa0eb6d" gracePeriod=30 Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.956096 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7684dfd8d9-wtn8h"] Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.963745 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.970901 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b251d66-e2c5-43ab-81db-339b0690d2dc-horizon-secret-key\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.970945 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b251d66-e2c5-43ab-81db-339b0690d2dc-logs\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.970998 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bhl5\" (UniqueName: \"kubernetes.io/projected/7b251d66-e2c5-43ab-81db-339b0690d2dc-kube-api-access-9bhl5\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.971043 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-scripts\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:11 crc kubenswrapper[4871]: I1007 23:51:11.971089 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-config-data\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.005544 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7684dfd8d9-wtn8h"] Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.057148 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.057373 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerName="glance-log" containerID="cri-o://6af6940675088239622d7ba305cd96b1b2662bf063f2be6efe35de6f0e8243eb" gracePeriod=30 Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.057768 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerName="glance-httpd" containerID="cri-o://bb9a8d3c4ec333ea45b6256627673301f12cf3af1bfea730f986e0ec05b71905" gracePeriod=30 Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.072878 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-scripts\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.072943 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66682\" (UniqueName: \"kubernetes.io/projected/75705323-d7c5-4134-bd47-30b62f198c7c-kube-api-access-66682\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.072971 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-scripts\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.073023 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-config-data\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.073040 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-config-data\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.073070 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75705323-d7c5-4134-bd47-30b62f198c7c-logs\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.073094 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75705323-d7c5-4134-bd47-30b62f198c7c-horizon-secret-key\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.073149 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b251d66-e2c5-43ab-81db-339b0690d2dc-horizon-secret-key\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.073170 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b251d66-e2c5-43ab-81db-339b0690d2dc-logs\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.073210 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bhl5\" (UniqueName: \"kubernetes.io/projected/7b251d66-e2c5-43ab-81db-339b0690d2dc-kube-api-access-9bhl5\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.074547 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-scripts\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.075423 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-config-data\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.077157 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b251d66-e2c5-43ab-81db-339b0690d2dc-logs\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.092538 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b251d66-e2c5-43ab-81db-339b0690d2dc-horizon-secret-key\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.108595 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bhl5\" (UniqueName: \"kubernetes.io/projected/7b251d66-e2c5-43ab-81db-339b0690d2dc-kube-api-access-9bhl5\") pod \"horizon-d94b76c57-wjk4c\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.131276 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.183256 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-scripts\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.183632 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66682\" (UniqueName: \"kubernetes.io/projected/75705323-d7c5-4134-bd47-30b62f198c7c-kube-api-access-66682\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.183746 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-config-data\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.183810 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75705323-d7c5-4134-bd47-30b62f198c7c-logs\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.183848 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75705323-d7c5-4134-bd47-30b62f198c7c-horizon-secret-key\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.186972 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75705323-d7c5-4134-bd47-30b62f198c7c-horizon-secret-key\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.187453 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-scripts\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.188748 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-config-data\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.188987 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75705323-d7c5-4134-bd47-30b62f198c7c-logs\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.218024 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66682\" (UniqueName: \"kubernetes.io/projected/75705323-d7c5-4134-bd47-30b62f198c7c-kube-api-access-66682\") pod \"horizon-7684dfd8d9-wtn8h\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.224162 4871 generic.go:334] "Generic (PLEG): container finished" podID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerID="a19d16bbb1da455b8bd1a86f2aecf55b9b15a8bfd8d48b6c2852c587c419ac82" exitCode=143 Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.224271 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e72ca756-e9c6-47f6-9a56-ea8f8da10938","Type":"ContainerDied","Data":"a19d16bbb1da455b8bd1a86f2aecf55b9b15a8bfd8d48b6c2852c587c419ac82"} Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.296637 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.441440 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7684dfd8d9-wtn8h"] Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.479832 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-668c994789-hrcgh"] Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.481751 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.500627 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-668c994789-hrcgh"] Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.593919 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kn42\" (UniqueName: \"kubernetes.io/projected/dca91af8-6003-4f54-ba25-d85809b9da33-kube-api-access-6kn42\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.594136 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dca91af8-6003-4f54-ba25-d85809b9da33-logs\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.594292 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-config-data\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.594382 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-scripts\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.594533 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dca91af8-6003-4f54-ba25-d85809b9da33-horizon-secret-key\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.640685 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d94b76c57-wjk4c"] Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.695803 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dca91af8-6003-4f54-ba25-d85809b9da33-horizon-secret-key\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.696120 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kn42\" (UniqueName: \"kubernetes.io/projected/dca91af8-6003-4f54-ba25-d85809b9da33-kube-api-access-6kn42\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.696143 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dca91af8-6003-4f54-ba25-d85809b9da33-logs\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.696252 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-config-data\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.696276 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-scripts\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.696763 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dca91af8-6003-4f54-ba25-d85809b9da33-logs\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.697212 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-scripts\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.697682 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-config-data\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.702022 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dca91af8-6003-4f54-ba25-d85809b9da33-horizon-secret-key\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.711656 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kn42\" (UniqueName: \"kubernetes.io/projected/dca91af8-6003-4f54-ba25-d85809b9da33-kube-api-access-6kn42\") pod \"horizon-668c994789-hrcgh\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.830372 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7684dfd8d9-wtn8h"] Oct 07 23:51:12 crc kubenswrapper[4871]: I1007 23:51:12.833393 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:12 crc kubenswrapper[4871]: W1007 23:51:12.833685 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75705323_d7c5_4134_bd47_30b62f198c7c.slice/crio-cbcfbb7004df1ccaccbad2fabb4db359b4cb507b60b648bbb22e46af58b3d4c1 WatchSource:0}: Error finding container cbcfbb7004df1ccaccbad2fabb4db359b4cb507b60b648bbb22e46af58b3d4c1: Status 404 returned error can't find the container with id cbcfbb7004df1ccaccbad2fabb4db359b4cb507b60b648bbb22e46af58b3d4c1 Oct 07 23:51:13 crc kubenswrapper[4871]: I1007 23:51:13.235900 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d94b76c57-wjk4c" event={"ID":"7b251d66-e2c5-43ab-81db-339b0690d2dc","Type":"ContainerStarted","Data":"778bfb8ed89e1024f295f7d949aee738f5446bc76974dadb9e2680014349b136"} Oct 07 23:51:13 crc kubenswrapper[4871]: I1007 23:51:13.239221 4871 generic.go:334] "Generic (PLEG): container finished" podID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerID="6af6940675088239622d7ba305cd96b1b2662bf063f2be6efe35de6f0e8243eb" exitCode=143 Oct 07 23:51:13 crc kubenswrapper[4871]: I1007 23:51:13.239293 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81394084-1cfb-4031-8c08-a7d59bf4d796","Type":"ContainerDied","Data":"6af6940675088239622d7ba305cd96b1b2662bf063f2be6efe35de6f0e8243eb"} Oct 07 23:51:13 crc kubenswrapper[4871]: I1007 23:51:13.240917 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7684dfd8d9-wtn8h" event={"ID":"75705323-d7c5-4134-bd47-30b62f198c7c","Type":"ContainerStarted","Data":"cbcfbb7004df1ccaccbad2fabb4db359b4cb507b60b648bbb22e46af58b3d4c1"} Oct 07 23:51:13 crc kubenswrapper[4871]: I1007 23:51:13.304877 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-668c994789-hrcgh"] Oct 07 23:51:13 crc kubenswrapper[4871]: W1007 23:51:13.310281 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddca91af8_6003_4f54_ba25_d85809b9da33.slice/crio-be8d11dca13a4ce476cdf66892063681825c5fca331165ce9ade3a2729f3c2fa WatchSource:0}: Error finding container be8d11dca13a4ce476cdf66892063681825c5fca331165ce9ade3a2729f3c2fa: Status 404 returned error can't find the container with id be8d11dca13a4ce476cdf66892063681825c5fca331165ce9ade3a2729f3c2fa Oct 07 23:51:14 crc kubenswrapper[4871]: I1007 23:51:14.250647 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-668c994789-hrcgh" event={"ID":"dca91af8-6003-4f54-ba25-d85809b9da33","Type":"ContainerStarted","Data":"be8d11dca13a4ce476cdf66892063681825c5fca331165ce9ade3a2729f3c2fa"} Oct 07 23:51:15 crc kubenswrapper[4871]: I1007 23:51:15.048228 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-74vwd"] Oct 07 23:51:15 crc kubenswrapper[4871]: I1007 23:51:15.059399 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-74vwd"] Oct 07 23:51:16 crc kubenswrapper[4871]: I1007 23:51:16.272693 4871 generic.go:334] "Generic (PLEG): container finished" podID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerID="d0811bd3822ca6ebeb120cdc6461892d29a96b4c5c061b4fc658953ddaa0eb6d" exitCode=0 Oct 07 23:51:16 crc kubenswrapper[4871]: I1007 23:51:16.272798 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e72ca756-e9c6-47f6-9a56-ea8f8da10938","Type":"ContainerDied","Data":"d0811bd3822ca6ebeb120cdc6461892d29a96b4c5c061b4fc658953ddaa0eb6d"} Oct 07 23:51:16 crc kubenswrapper[4871]: I1007 23:51:16.275056 4871 generic.go:334] "Generic (PLEG): container finished" podID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerID="bb9a8d3c4ec333ea45b6256627673301f12cf3af1bfea730f986e0ec05b71905" exitCode=0 Oct 07 23:51:16 crc kubenswrapper[4871]: I1007 23:51:16.275097 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81394084-1cfb-4031-8c08-a7d59bf4d796","Type":"ContainerDied","Data":"bb9a8d3c4ec333ea45b6256627673301f12cf3af1bfea730f986e0ec05b71905"} Oct 07 23:51:16 crc kubenswrapper[4871]: I1007 23:51:16.998104 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed609edd-a961-4533-a12d-11adc1800513" path="/var/lib/kubelet/pods/ed609edd-a961-4533-a12d-11adc1800513/volumes" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.653211 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.779648 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-ceph\") pod \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.779783 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-scripts\") pod \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.779916 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5fv6\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-kube-api-access-k5fv6\") pod \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.779947 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-combined-ca-bundle\") pod \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.780034 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-httpd-run\") pod \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.780090 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-config-data\") pod \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.780119 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-logs\") pod \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\" (UID: \"e72ca756-e9c6-47f6-9a56-ea8f8da10938\") " Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.780426 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e72ca756-e9c6-47f6-9a56-ea8f8da10938" (UID: "e72ca756-e9c6-47f6-9a56-ea8f8da10938"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.780633 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.780769 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-logs" (OuterVolumeSpecName: "logs") pod "e72ca756-e9c6-47f6-9a56-ea8f8da10938" (UID: "e72ca756-e9c6-47f6-9a56-ea8f8da10938"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.783681 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-scripts" (OuterVolumeSpecName: "scripts") pod "e72ca756-e9c6-47f6-9a56-ea8f8da10938" (UID: "e72ca756-e9c6-47f6-9a56-ea8f8da10938"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.784432 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-kube-api-access-k5fv6" (OuterVolumeSpecName: "kube-api-access-k5fv6") pod "e72ca756-e9c6-47f6-9a56-ea8f8da10938" (UID: "e72ca756-e9c6-47f6-9a56-ea8f8da10938"). InnerVolumeSpecName "kube-api-access-k5fv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.784976 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-ceph" (OuterVolumeSpecName: "ceph") pod "e72ca756-e9c6-47f6-9a56-ea8f8da10938" (UID: "e72ca756-e9c6-47f6-9a56-ea8f8da10938"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.863842 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e72ca756-e9c6-47f6-9a56-ea8f8da10938" (UID: "e72ca756-e9c6-47f6-9a56-ea8f8da10938"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.882522 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.882555 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.882568 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5fv6\" (UniqueName: \"kubernetes.io/projected/e72ca756-e9c6-47f6-9a56-ea8f8da10938-kube-api-access-k5fv6\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.882582 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.882592 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e72ca756-e9c6-47f6-9a56-ea8f8da10938-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.891975 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-config-data" (OuterVolumeSpecName: "config-data") pod "e72ca756-e9c6-47f6-9a56-ea8f8da10938" (UID: "e72ca756-e9c6-47f6-9a56-ea8f8da10938"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:51:20 crc kubenswrapper[4871]: I1007 23:51:20.984530 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e72ca756-e9c6-47f6-9a56-ea8f8da10938-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.143368 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.288569 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-scripts\") pod \"81394084-1cfb-4031-8c08-a7d59bf4d796\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.288730 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-httpd-run\") pod \"81394084-1cfb-4031-8c08-a7d59bf4d796\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.288832 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-ceph\") pod \"81394084-1cfb-4031-8c08-a7d59bf4d796\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.288907 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfw7v\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-kube-api-access-nfw7v\") pod \"81394084-1cfb-4031-8c08-a7d59bf4d796\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.288947 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-logs\") pod \"81394084-1cfb-4031-8c08-a7d59bf4d796\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.289151 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-config-data\") pod \"81394084-1cfb-4031-8c08-a7d59bf4d796\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.289233 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-combined-ca-bundle\") pod \"81394084-1cfb-4031-8c08-a7d59bf4d796\" (UID: \"81394084-1cfb-4031-8c08-a7d59bf4d796\") " Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.289235 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "81394084-1cfb-4031-8c08-a7d59bf4d796" (UID: "81394084-1cfb-4031-8c08-a7d59bf4d796"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.289518 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-logs" (OuterVolumeSpecName: "logs") pod "81394084-1cfb-4031-8c08-a7d59bf4d796" (UID: "81394084-1cfb-4031-8c08-a7d59bf4d796"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.294509 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-kube-api-access-nfw7v" (OuterVolumeSpecName: "kube-api-access-nfw7v") pod "81394084-1cfb-4031-8c08-a7d59bf4d796" (UID: "81394084-1cfb-4031-8c08-a7d59bf4d796"). InnerVolumeSpecName "kube-api-access-nfw7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.294512 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-scripts" (OuterVolumeSpecName: "scripts") pod "81394084-1cfb-4031-8c08-a7d59bf4d796" (UID: "81394084-1cfb-4031-8c08-a7d59bf4d796"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.299369 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.299411 4871 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.299424 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfw7v\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-kube-api-access-nfw7v\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.299437 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81394084-1cfb-4031-8c08-a7d59bf4d796-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.309529 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-ceph" (OuterVolumeSpecName: "ceph") pod "81394084-1cfb-4031-8c08-a7d59bf4d796" (UID: "81394084-1cfb-4031-8c08-a7d59bf4d796"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.331074 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.331101 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81394084-1cfb-4031-8c08-a7d59bf4d796","Type":"ContainerDied","Data":"4daa10b0354e0d7d7b646b7b051440659d228c97920a1a5aa02b179937a41ec5"} Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.331167 4871 scope.go:117] "RemoveContainer" containerID="bb9a8d3c4ec333ea45b6256627673301f12cf3af1bfea730f986e0ec05b71905" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.339010 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-668c994789-hrcgh" event={"ID":"dca91af8-6003-4f54-ba25-d85809b9da33","Type":"ContainerStarted","Data":"713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc"} Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.339054 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-668c994789-hrcgh" event={"ID":"dca91af8-6003-4f54-ba25-d85809b9da33","Type":"ContainerStarted","Data":"a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f"} Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.351457 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81394084-1cfb-4031-8c08-a7d59bf4d796" (UID: "81394084-1cfb-4031-8c08-a7d59bf4d796"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.358858 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7684dfd8d9-wtn8h" event={"ID":"75705323-d7c5-4134-bd47-30b62f198c7c","Type":"ContainerStarted","Data":"be2220e3e81b52aea4b4acea3c4e239422cfa92c76c97b51e7e8cd27e23a74eb"} Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.358899 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7684dfd8d9-wtn8h" event={"ID":"75705323-d7c5-4134-bd47-30b62f198c7c","Type":"ContainerStarted","Data":"2181a3fb92482fe53a2cc9f8a4e63b14fd7608e47005c74226f5653ddc691f84"} Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.359038 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7684dfd8d9-wtn8h" podUID="75705323-d7c5-4134-bd47-30b62f198c7c" containerName="horizon-log" containerID="cri-o://2181a3fb92482fe53a2cc9f8a4e63b14fd7608e47005c74226f5653ddc691f84" gracePeriod=30 Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.359091 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7684dfd8d9-wtn8h" podUID="75705323-d7c5-4134-bd47-30b62f198c7c" containerName="horizon" containerID="cri-o://be2220e3e81b52aea4b4acea3c4e239422cfa92c76c97b51e7e8cd27e23a74eb" gracePeriod=30 Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.372632 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d94b76c57-wjk4c" event={"ID":"7b251d66-e2c5-43ab-81db-339b0690d2dc","Type":"ContainerStarted","Data":"74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626"} Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.372672 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d94b76c57-wjk4c" event={"ID":"7b251d66-e2c5-43ab-81db-339b0690d2dc","Type":"ContainerStarted","Data":"cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad"} Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.376919 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-config-data" (OuterVolumeSpecName: "config-data") pod "81394084-1cfb-4031-8c08-a7d59bf4d796" (UID: "81394084-1cfb-4031-8c08-a7d59bf4d796"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.378385 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e72ca756-e9c6-47f6-9a56-ea8f8da10938","Type":"ContainerDied","Data":"68e8bcbb6e10488cd59342a92b7fe7296e7c928fb4aa6d5c1d1a3878dff9b9c1"} Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.378470 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.392646 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-668c994789-hrcgh" podStartSLOduration=2.396013429 podStartE2EDuration="9.392626325s" podCreationTimestamp="2025-10-07 23:51:12 +0000 UTC" firstStartedPulling="2025-10-07 23:51:13.318508266 +0000 UTC m=+6147.121206349" lastFinishedPulling="2025-10-07 23:51:20.315121172 +0000 UTC m=+6154.117819245" observedRunningTime="2025-10-07 23:51:21.367564087 +0000 UTC m=+6155.170262160" watchObservedRunningTime="2025-10-07 23:51:21.392626325 +0000 UTC m=+6155.195324398" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.398374 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7684dfd8d9-wtn8h" podStartSLOduration=2.910639318 podStartE2EDuration="10.398356585s" podCreationTimestamp="2025-10-07 23:51:11 +0000 UTC" firstStartedPulling="2025-10-07 23:51:12.846839535 +0000 UTC m=+6146.649537608" lastFinishedPulling="2025-10-07 23:51:20.334556802 +0000 UTC m=+6154.137254875" observedRunningTime="2025-10-07 23:51:21.385085677 +0000 UTC m=+6155.187783750" watchObservedRunningTime="2025-10-07 23:51:21.398356585 +0000 UTC m=+6155.201054658" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.404279 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.404530 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81394084-1cfb-4031-8c08-a7d59bf4d796-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.404584 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/81394084-1cfb-4031-8c08-a7d59bf4d796-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.406053 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-d94b76c57-wjk4c" podStartSLOduration=2.7511372339999998 podStartE2EDuration="10.406035846s" podCreationTimestamp="2025-10-07 23:51:11 +0000 UTC" firstStartedPulling="2025-10-07 23:51:12.659263075 +0000 UTC m=+6146.461961138" lastFinishedPulling="2025-10-07 23:51:20.314161677 +0000 UTC m=+6154.116859750" observedRunningTime="2025-10-07 23:51:21.403363046 +0000 UTC m=+6155.206061119" watchObservedRunningTime="2025-10-07 23:51:21.406035846 +0000 UTC m=+6155.208733919" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.448264 4871 scope.go:117] "RemoveContainer" containerID="6af6940675088239622d7ba305cd96b1b2662bf063f2be6efe35de6f0e8243eb" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.451697 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.459151 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.472462 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:51:21 crc kubenswrapper[4871]: E1007 23:51:21.473025 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerName="glance-log" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.473093 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerName="glance-log" Oct 07 23:51:21 crc kubenswrapper[4871]: E1007 23:51:21.473165 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerName="glance-log" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.473212 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerName="glance-log" Oct 07 23:51:21 crc kubenswrapper[4871]: E1007 23:51:21.473270 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerName="glance-httpd" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.475954 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerName="glance-httpd" Oct 07 23:51:21 crc kubenswrapper[4871]: E1007 23:51:21.476009 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerName="glance-httpd" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.476062 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerName="glance-httpd" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.476277 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerName="glance-log" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.476330 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" containerName="glance-httpd" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.476388 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerName="glance-httpd" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.476434 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="81394084-1cfb-4031-8c08-a7d59bf4d796" containerName="glance-log" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.473105 4871 scope.go:117] "RemoveContainer" containerID="d0811bd3822ca6ebeb120cdc6461892d29a96b4c5c061b4fc658953ddaa0eb6d" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.477641 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.481280 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.493944 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.528642 4871 scope.go:117] "RemoveContainer" containerID="a19d16bbb1da455b8bd1a86f2aecf55b9b15a8bfd8d48b6c2852c587c419ac82" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.608878 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02e76f14-d4ce-486a-8979-ff8a12871783-scripts\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.609147 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e76f14-d4ce-486a-8979-ff8a12871783-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.609178 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02e76f14-d4ce-486a-8979-ff8a12871783-config-data\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.609202 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/02e76f14-d4ce-486a-8979-ff8a12871783-ceph\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.609235 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02e76f14-d4ce-486a-8979-ff8a12871783-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.609286 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxxsv\" (UniqueName: \"kubernetes.io/projected/02e76f14-d4ce-486a-8979-ff8a12871783-kube-api-access-rxxsv\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.609331 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02e76f14-d4ce-486a-8979-ff8a12871783-logs\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.669896 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.678109 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.706830 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.708513 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.711228 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02e76f14-d4ce-486a-8979-ff8a12871783-logs\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.711319 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02e76f14-d4ce-486a-8979-ff8a12871783-scripts\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.711351 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e76f14-d4ce-486a-8979-ff8a12871783-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.711366 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02e76f14-d4ce-486a-8979-ff8a12871783-config-data\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.711390 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/02e76f14-d4ce-486a-8979-ff8a12871783-ceph\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.711419 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02e76f14-d4ce-486a-8979-ff8a12871783-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.711482 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxxsv\" (UniqueName: \"kubernetes.io/projected/02e76f14-d4ce-486a-8979-ff8a12871783-kube-api-access-rxxsv\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.712296 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02e76f14-d4ce-486a-8979-ff8a12871783-logs\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.715666 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.721642 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/02e76f14-d4ce-486a-8979-ff8a12871783-ceph\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.722051 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/02e76f14-d4ce-486a-8979-ff8a12871783-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.724352 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.728359 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02e76f14-d4ce-486a-8979-ff8a12871783-config-data\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.730071 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02e76f14-d4ce-486a-8979-ff8a12871783-scripts\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.730282 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e76f14-d4ce-486a-8979-ff8a12871783-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.734201 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxxsv\" (UniqueName: \"kubernetes.io/projected/02e76f14-d4ce-486a-8979-ff8a12871783-kube-api-access-rxxsv\") pod \"glance-default-external-api-0\" (UID: \"02e76f14-d4ce-486a-8979-ff8a12871783\") " pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.812374 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.812995 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c54d4124-f070-4e87-bfd7-6f319c374d36-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.813072 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgkgs\" (UniqueName: \"kubernetes.io/projected/c54d4124-f070-4e87-bfd7-6f319c374d36-kube-api-access-mgkgs\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.813092 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c54d4124-f070-4e87-bfd7-6f319c374d36-logs\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.813147 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c54d4124-f070-4e87-bfd7-6f319c374d36-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.813228 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c54d4124-f070-4e87-bfd7-6f319c374d36-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.813315 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c54d4124-f070-4e87-bfd7-6f319c374d36-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.813354 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c54d4124-f070-4e87-bfd7-6f319c374d36-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.918400 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgkgs\" (UniqueName: \"kubernetes.io/projected/c54d4124-f070-4e87-bfd7-6f319c374d36-kube-api-access-mgkgs\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.918438 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c54d4124-f070-4e87-bfd7-6f319c374d36-logs\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.918469 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c54d4124-f070-4e87-bfd7-6f319c374d36-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.918539 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c54d4124-f070-4e87-bfd7-6f319c374d36-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.918595 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c54d4124-f070-4e87-bfd7-6f319c374d36-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.918633 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c54d4124-f070-4e87-bfd7-6f319c374d36-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.918662 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c54d4124-f070-4e87-bfd7-6f319c374d36-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.924355 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c54d4124-f070-4e87-bfd7-6f319c374d36-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.924627 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c54d4124-f070-4e87-bfd7-6f319c374d36-logs\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.924632 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c54d4124-f070-4e87-bfd7-6f319c374d36-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.927487 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c54d4124-f070-4e87-bfd7-6f319c374d36-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.929451 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c54d4124-f070-4e87-bfd7-6f319c374d36-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.938678 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c54d4124-f070-4e87-bfd7-6f319c374d36-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:21 crc kubenswrapper[4871]: I1007 23:51:21.939459 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgkgs\" (UniqueName: \"kubernetes.io/projected/c54d4124-f070-4e87-bfd7-6f319c374d36-kube-api-access-mgkgs\") pod \"glance-default-internal-api-0\" (UID: \"c54d4124-f070-4e87-bfd7-6f319c374d36\") " pod="openstack/glance-default-internal-api-0" Oct 07 23:51:22 crc kubenswrapper[4871]: I1007 23:51:22.085922 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:22 crc kubenswrapper[4871]: I1007 23:51:22.133199 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:22 crc kubenswrapper[4871]: I1007 23:51:22.133232 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:22 crc kubenswrapper[4871]: I1007 23:51:22.298003 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:22 crc kubenswrapper[4871]: I1007 23:51:22.379995 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 23:51:22 crc kubenswrapper[4871]: I1007 23:51:22.392431 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02e76f14-d4ce-486a-8979-ff8a12871783","Type":"ContainerStarted","Data":"3194fa4b564f55cf5da7ee5f178a1261d1ce82a7185983ab244caf9f42882389"} Oct 07 23:51:22 crc kubenswrapper[4871]: I1007 23:51:22.590092 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 23:51:22 crc kubenswrapper[4871]: I1007 23:51:22.833491 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:22 crc kubenswrapper[4871]: I1007 23:51:22.833728 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:23 crc kubenswrapper[4871]: I1007 23:51:23.000179 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81394084-1cfb-4031-8c08-a7d59bf4d796" path="/var/lib/kubelet/pods/81394084-1cfb-4031-8c08-a7d59bf4d796/volumes" Oct 07 23:51:23 crc kubenswrapper[4871]: I1007 23:51:23.001011 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e72ca756-e9c6-47f6-9a56-ea8f8da10938" path="/var/lib/kubelet/pods/e72ca756-e9c6-47f6-9a56-ea8f8da10938/volumes" Oct 07 23:51:23 crc kubenswrapper[4871]: I1007 23:51:23.402635 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02e76f14-d4ce-486a-8979-ff8a12871783","Type":"ContainerStarted","Data":"de329b485359b424bb3f46fe8ece8bf8d78c9019c268687b0d4030e3edca4446"} Oct 07 23:51:23 crc kubenswrapper[4871]: I1007 23:51:23.405924 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c54d4124-f070-4e87-bfd7-6f319c374d36","Type":"ContainerStarted","Data":"5d370b7be0eb00c69640f3fbac24b06472828f7a57a03e38f995522baa607e13"} Oct 07 23:51:23 crc kubenswrapper[4871]: I1007 23:51:23.406012 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c54d4124-f070-4e87-bfd7-6f319c374d36","Type":"ContainerStarted","Data":"879050aa40fa51689e60065f8654c2b0e821b7602c06e33a2c9e8eb918bdb246"} Oct 07 23:51:24 crc kubenswrapper[4871]: I1007 23:51:24.415775 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c54d4124-f070-4e87-bfd7-6f319c374d36","Type":"ContainerStarted","Data":"2c4905a1206932ff9b5dde5df87be6fbcde563c968570499d52e27869af010b4"} Oct 07 23:51:24 crc kubenswrapper[4871]: I1007 23:51:24.417618 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"02e76f14-d4ce-486a-8979-ff8a12871783","Type":"ContainerStarted","Data":"48a122819b8b49e88f84fcba0a871fd2323b26e762dec418e2cd30a458eaa922"} Oct 07 23:51:24 crc kubenswrapper[4871]: I1007 23:51:24.438653 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.438635369 podStartE2EDuration="3.438635369s" podCreationTimestamp="2025-10-07 23:51:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:51:24.436422961 +0000 UTC m=+6158.239121044" watchObservedRunningTime="2025-10-07 23:51:24.438635369 +0000 UTC m=+6158.241333462" Oct 07 23:51:24 crc kubenswrapper[4871]: I1007 23:51:24.466104 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.466060918 podStartE2EDuration="3.466060918s" podCreationTimestamp="2025-10-07 23:51:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:51:24.462086084 +0000 UTC m=+6158.264784157" watchObservedRunningTime="2025-10-07 23:51:24.466060918 +0000 UTC m=+6158.268759001" Oct 07 23:51:26 crc kubenswrapper[4871]: I1007 23:51:26.022772 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-557c-account-create-8mc29"] Oct 07 23:51:26 crc kubenswrapper[4871]: I1007 23:51:26.030021 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-557c-account-create-8mc29"] Oct 07 23:51:26 crc kubenswrapper[4871]: I1007 23:51:26.998648 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbc912d0-51de-4f35-85b3-2f3403d2c554" path="/var/lib/kubelet/pods/cbc912d0-51de-4f35-85b3-2f3403d2c554/volumes" Oct 07 23:51:31 crc kubenswrapper[4871]: I1007 23:51:31.812920 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 23:51:31 crc kubenswrapper[4871]: I1007 23:51:31.813960 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 23:51:31 crc kubenswrapper[4871]: I1007 23:51:31.856920 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 23:51:31 crc kubenswrapper[4871]: I1007 23:51:31.894766 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.086246 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.086550 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.116736 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.134828 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-d94b76c57-wjk4c" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.153085 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.535776 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.536196 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.536220 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.536238 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 23:51:32 crc kubenswrapper[4871]: I1007 23:51:32.835171 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-668c994789-hrcgh" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Oct 07 23:51:34 crc kubenswrapper[4871]: I1007 23:51:34.033337 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-7vdvf"] Oct 07 23:51:34 crc kubenswrapper[4871]: I1007 23:51:34.042769 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-7vdvf"] Oct 07 23:51:34 crc kubenswrapper[4871]: I1007 23:51:34.554918 4871 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 23:51:34 crc kubenswrapper[4871]: I1007 23:51:34.554945 4871 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 23:51:34 crc kubenswrapper[4871]: I1007 23:51:34.611449 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 23:51:34 crc kubenswrapper[4871]: I1007 23:51:34.611740 4871 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 23:51:34 crc kubenswrapper[4871]: I1007 23:51:34.810289 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:34 crc kubenswrapper[4871]: I1007 23:51:34.882626 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 23:51:34 crc kubenswrapper[4871]: I1007 23:51:34.992651 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b06d89-dec9-4efd-9e84-be7b357e916a" path="/var/lib/kubelet/pods/f4b06d89-dec9-4efd-9e84-be7b357e916a/volumes" Oct 07 23:51:35 crc kubenswrapper[4871]: I1007 23:51:35.134878 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 23:51:35 crc kubenswrapper[4871]: I1007 23:51:35.511906 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:51:35 crc kubenswrapper[4871]: I1007 23:51:35.512276 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:51:43 crc kubenswrapper[4871]: I1007 23:51:43.917588 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:44 crc kubenswrapper[4871]: I1007 23:51:44.643506 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:45 crc kubenswrapper[4871]: I1007 23:51:45.377876 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:51:46 crc kubenswrapper[4871]: I1007 23:51:46.380546 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:51:46 crc kubenswrapper[4871]: I1007 23:51:46.440841 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d94b76c57-wjk4c"] Oct 07 23:51:46 crc kubenswrapper[4871]: I1007 23:51:46.441046 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d94b76c57-wjk4c" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon-log" containerID="cri-o://cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad" gracePeriod=30 Oct 07 23:51:46 crc kubenswrapper[4871]: I1007 23:51:46.441448 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d94b76c57-wjk4c" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon" containerID="cri-o://74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626" gracePeriod=30 Oct 07 23:51:49 crc kubenswrapper[4871]: I1007 23:51:49.709569 4871 generic.go:334] "Generic (PLEG): container finished" podID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerID="74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626" exitCode=0 Oct 07 23:51:49 crc kubenswrapper[4871]: I1007 23:51:49.709622 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d94b76c57-wjk4c" event={"ID":"7b251d66-e2c5-43ab-81db-339b0690d2dc","Type":"ContainerDied","Data":"74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626"} Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.730089 4871 generic.go:334] "Generic (PLEG): container finished" podID="75705323-d7c5-4134-bd47-30b62f198c7c" containerID="be2220e3e81b52aea4b4acea3c4e239422cfa92c76c97b51e7e8cd27e23a74eb" exitCode=137 Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.730636 4871 generic.go:334] "Generic (PLEG): container finished" podID="75705323-d7c5-4134-bd47-30b62f198c7c" containerID="2181a3fb92482fe53a2cc9f8a4e63b14fd7608e47005c74226f5653ddc691f84" exitCode=137 Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.730162 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7684dfd8d9-wtn8h" event={"ID":"75705323-d7c5-4134-bd47-30b62f198c7c","Type":"ContainerDied","Data":"be2220e3e81b52aea4b4acea3c4e239422cfa92c76c97b51e7e8cd27e23a74eb"} Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.730670 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7684dfd8d9-wtn8h" event={"ID":"75705323-d7c5-4134-bd47-30b62f198c7c","Type":"ContainerDied","Data":"2181a3fb92482fe53a2cc9f8a4e63b14fd7608e47005c74226f5653ddc691f84"} Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.827674 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.931979 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66682\" (UniqueName: \"kubernetes.io/projected/75705323-d7c5-4134-bd47-30b62f198c7c-kube-api-access-66682\") pod \"75705323-d7c5-4134-bd47-30b62f198c7c\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.932119 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75705323-d7c5-4134-bd47-30b62f198c7c-logs\") pod \"75705323-d7c5-4134-bd47-30b62f198c7c\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.932186 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-scripts\") pod \"75705323-d7c5-4134-bd47-30b62f198c7c\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.932215 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-config-data\") pod \"75705323-d7c5-4134-bd47-30b62f198c7c\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.932272 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75705323-d7c5-4134-bd47-30b62f198c7c-horizon-secret-key\") pod \"75705323-d7c5-4134-bd47-30b62f198c7c\" (UID: \"75705323-d7c5-4134-bd47-30b62f198c7c\") " Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.932629 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75705323-d7c5-4134-bd47-30b62f198c7c-logs" (OuterVolumeSpecName: "logs") pod "75705323-d7c5-4134-bd47-30b62f198c7c" (UID: "75705323-d7c5-4134-bd47-30b62f198c7c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.933065 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75705323-d7c5-4134-bd47-30b62f198c7c-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.937944 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75705323-d7c5-4134-bd47-30b62f198c7c-kube-api-access-66682" (OuterVolumeSpecName: "kube-api-access-66682") pod "75705323-d7c5-4134-bd47-30b62f198c7c" (UID: "75705323-d7c5-4134-bd47-30b62f198c7c"). InnerVolumeSpecName "kube-api-access-66682". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.939548 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75705323-d7c5-4134-bd47-30b62f198c7c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "75705323-d7c5-4134-bd47-30b62f198c7c" (UID: "75705323-d7c5-4134-bd47-30b62f198c7c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.959124 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-scripts" (OuterVolumeSpecName: "scripts") pod "75705323-d7c5-4134-bd47-30b62f198c7c" (UID: "75705323-d7c5-4134-bd47-30b62f198c7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:51:51 crc kubenswrapper[4871]: I1007 23:51:51.972447 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-config-data" (OuterVolumeSpecName: "config-data") pod "75705323-d7c5-4134-bd47-30b62f198c7c" (UID: "75705323-d7c5-4134-bd47-30b62f198c7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.035530 4871 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75705323-d7c5-4134-bd47-30b62f198c7c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.035583 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66682\" (UniqueName: \"kubernetes.io/projected/75705323-d7c5-4134-bd47-30b62f198c7c-kube-api-access-66682\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.035622 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.035641 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75705323-d7c5-4134-bd47-30b62f198c7c-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.133214 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d94b76c57-wjk4c" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.743051 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7684dfd8d9-wtn8h" event={"ID":"75705323-d7c5-4134-bd47-30b62f198c7c","Type":"ContainerDied","Data":"cbcfbb7004df1ccaccbad2fabb4db359b4cb507b60b648bbb22e46af58b3d4c1"} Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.744211 4871 scope.go:117] "RemoveContainer" containerID="be2220e3e81b52aea4b4acea3c4e239422cfa92c76c97b51e7e8cd27e23a74eb" Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.744487 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7684dfd8d9-wtn8h" Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.818212 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7684dfd8d9-wtn8h"] Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.830856 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7684dfd8d9-wtn8h"] Oct 07 23:51:52 crc kubenswrapper[4871]: I1007 23:51:52.965142 4871 scope.go:117] "RemoveContainer" containerID="2181a3fb92482fe53a2cc9f8a4e63b14fd7608e47005c74226f5653ddc691f84" Oct 07 23:51:53 crc kubenswrapper[4871]: I1007 23:51:53.006021 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75705323-d7c5-4134-bd47-30b62f198c7c" path="/var/lib/kubelet/pods/75705323-d7c5-4134-bd47-30b62f198c7c/volumes" Oct 07 23:51:59 crc kubenswrapper[4871]: I1007 23:51:59.001629 4871 scope.go:117] "RemoveContainer" containerID="67eaf865e8d6d15aacd921116b15cab4bc1d0e2fe81329a467310321e0c57244" Oct 07 23:51:59 crc kubenswrapper[4871]: I1007 23:51:59.033923 4871 scope.go:117] "RemoveContainer" containerID="7453dba567a5dcc87ebe1c6e53a79f688d273ce2f7a2981e5d483e37fe9027b3" Oct 07 23:51:59 crc kubenswrapper[4871]: I1007 23:51:59.111170 4871 scope.go:117] "RemoveContainer" containerID="f9baa490831c655023bad085dd44139134fa574578421a343dbdf99605b825d2" Oct 07 23:52:02 crc kubenswrapper[4871]: I1007 23:52:02.132817 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d94b76c57-wjk4c" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 07 23:52:04 crc kubenswrapper[4871]: I1007 23:52:04.064703 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wws9x"] Oct 07 23:52:04 crc kubenswrapper[4871]: I1007 23:52:04.077771 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wws9x"] Oct 07 23:52:04 crc kubenswrapper[4871]: I1007 23:52:04.997922 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11b57069-5879-4ffe-9574-70103d3e0f58" path="/var/lib/kubelet/pods/11b57069-5879-4ffe-9574-70103d3e0f58/volumes" Oct 07 23:52:05 crc kubenswrapper[4871]: I1007 23:52:05.513033 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:52:05 crc kubenswrapper[4871]: I1007 23:52:05.513124 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:52:05 crc kubenswrapper[4871]: I1007 23:52:05.513180 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:52:05 crc kubenswrapper[4871]: I1007 23:52:05.514908 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cee94e818dbd839f5c342b9e271035e3e73be566c36380410bb63a203783329e"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:52:05 crc kubenswrapper[4871]: I1007 23:52:05.515039 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://cee94e818dbd839f5c342b9e271035e3e73be566c36380410bb63a203783329e" gracePeriod=600 Oct 07 23:52:05 crc kubenswrapper[4871]: I1007 23:52:05.901047 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="cee94e818dbd839f5c342b9e271035e3e73be566c36380410bb63a203783329e" exitCode=0 Oct 07 23:52:05 crc kubenswrapper[4871]: I1007 23:52:05.901221 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"cee94e818dbd839f5c342b9e271035e3e73be566c36380410bb63a203783329e"} Oct 07 23:52:05 crc kubenswrapper[4871]: I1007 23:52:05.901460 4871 scope.go:117] "RemoveContainer" containerID="abd5595703ae09449a5e42d2b4faca67aa152e790a8a2c7b191b268b2a8789c5" Oct 07 23:52:06 crc kubenswrapper[4871]: I1007 23:52:06.914872 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104"} Oct 07 23:52:12 crc kubenswrapper[4871]: I1007 23:52:12.133460 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d94b76c57-wjk4c" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 07 23:52:12 crc kubenswrapper[4871]: I1007 23:52:12.134383 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:52:14 crc kubenswrapper[4871]: I1007 23:52:14.076528 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-3b90-account-create-rdbvt"] Oct 07 23:52:14 crc kubenswrapper[4871]: I1007 23:52:14.095917 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-3b90-account-create-rdbvt"] Oct 07 23:52:14 crc kubenswrapper[4871]: I1007 23:52:14.998203 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5d3c2db-5aee-4904-8bc7-50c0b556eceb" path="/var/lib/kubelet/pods/b5d3c2db-5aee-4904-8bc7-50c0b556eceb/volumes" Oct 07 23:52:16 crc kubenswrapper[4871]: I1007 23:52:16.835927 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.014346 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-scripts\") pod \"7b251d66-e2c5-43ab-81db-339b0690d2dc\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.014549 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bhl5\" (UniqueName: \"kubernetes.io/projected/7b251d66-e2c5-43ab-81db-339b0690d2dc-kube-api-access-9bhl5\") pod \"7b251d66-e2c5-43ab-81db-339b0690d2dc\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.014586 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-config-data\") pod \"7b251d66-e2c5-43ab-81db-339b0690d2dc\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.014668 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b251d66-e2c5-43ab-81db-339b0690d2dc-logs\") pod \"7b251d66-e2c5-43ab-81db-339b0690d2dc\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.014689 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b251d66-e2c5-43ab-81db-339b0690d2dc-horizon-secret-key\") pod \"7b251d66-e2c5-43ab-81db-339b0690d2dc\" (UID: \"7b251d66-e2c5-43ab-81db-339b0690d2dc\") " Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.015417 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b251d66-e2c5-43ab-81db-339b0690d2dc-logs" (OuterVolumeSpecName: "logs") pod "7b251d66-e2c5-43ab-81db-339b0690d2dc" (UID: "7b251d66-e2c5-43ab-81db-339b0690d2dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.016100 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b251d66-e2c5-43ab-81db-339b0690d2dc-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.022066 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b251d66-e2c5-43ab-81db-339b0690d2dc-kube-api-access-9bhl5" (OuterVolumeSpecName: "kube-api-access-9bhl5") pod "7b251d66-e2c5-43ab-81db-339b0690d2dc" (UID: "7b251d66-e2c5-43ab-81db-339b0690d2dc"). InnerVolumeSpecName "kube-api-access-9bhl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.023446 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b251d66-e2c5-43ab-81db-339b0690d2dc-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7b251d66-e2c5-43ab-81db-339b0690d2dc" (UID: "7b251d66-e2c5-43ab-81db-339b0690d2dc"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.033062 4871 generic.go:334] "Generic (PLEG): container finished" podID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerID="cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad" exitCode=137 Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.033123 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d94b76c57-wjk4c" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.033134 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d94b76c57-wjk4c" event={"ID":"7b251d66-e2c5-43ab-81db-339b0690d2dc","Type":"ContainerDied","Data":"cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad"} Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.033203 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d94b76c57-wjk4c" event={"ID":"7b251d66-e2c5-43ab-81db-339b0690d2dc","Type":"ContainerDied","Data":"778bfb8ed89e1024f295f7d949aee738f5446bc76974dadb9e2680014349b136"} Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.033244 4871 scope.go:117] "RemoveContainer" containerID="74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.040741 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-config-data" (OuterVolumeSpecName: "config-data") pod "7b251d66-e2c5-43ab-81db-339b0690d2dc" (UID: "7b251d66-e2c5-43ab-81db-339b0690d2dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.041012 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-scripts" (OuterVolumeSpecName: "scripts") pod "7b251d66-e2c5-43ab-81db-339b0690d2dc" (UID: "7b251d66-e2c5-43ab-81db-339b0690d2dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.118897 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.118945 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bhl5\" (UniqueName: \"kubernetes.io/projected/7b251d66-e2c5-43ab-81db-339b0690d2dc-kube-api-access-9bhl5\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.118963 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7b251d66-e2c5-43ab-81db-339b0690d2dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.118982 4871 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7b251d66-e2c5-43ab-81db-339b0690d2dc-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.373619 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d94b76c57-wjk4c"] Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.374294 4871 scope.go:117] "RemoveContainer" containerID="cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.381962 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d94b76c57-wjk4c"] Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.396040 4871 scope.go:117] "RemoveContainer" containerID="74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626" Oct 07 23:52:17 crc kubenswrapper[4871]: E1007 23:52:17.396438 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626\": container with ID starting with 74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626 not found: ID does not exist" containerID="74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.396479 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626"} err="failed to get container status \"74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626\": rpc error: code = NotFound desc = could not find container \"74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626\": container with ID starting with 74b098d4ecac2e803ab4409061128e3e4653ffe3156ca3dd65fa77dacd5c1626 not found: ID does not exist" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.396511 4871 scope.go:117] "RemoveContainer" containerID="cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad" Oct 07 23:52:17 crc kubenswrapper[4871]: E1007 23:52:17.396778 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad\": container with ID starting with cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad not found: ID does not exist" containerID="cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad" Oct 07 23:52:17 crc kubenswrapper[4871]: I1007 23:52:17.396863 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad"} err="failed to get container status \"cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad\": rpc error: code = NotFound desc = could not find container \"cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad\": container with ID starting with cbd3590b47a0445e28c936b7cf68f3d593fd6e9a50ba878bde5f8a2a44cb68ad not found: ID does not exist" Oct 07 23:52:18 crc kubenswrapper[4871]: I1007 23:52:18.998636 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" path="/var/lib/kubelet/pods/7b251d66-e2c5-43ab-81db-339b0690d2dc/volumes" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.408734 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f455d5855-8bgr8"] Oct 07 23:52:20 crc kubenswrapper[4871]: E1007 23:52:20.409524 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon-log" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.409539 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon-log" Oct 07 23:52:20 crc kubenswrapper[4871]: E1007 23:52:20.409562 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75705323-d7c5-4134-bd47-30b62f198c7c" containerName="horizon-log" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.409570 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="75705323-d7c5-4134-bd47-30b62f198c7c" containerName="horizon-log" Oct 07 23:52:20 crc kubenswrapper[4871]: E1007 23:52:20.409597 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75705323-d7c5-4134-bd47-30b62f198c7c" containerName="horizon" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.409605 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="75705323-d7c5-4134-bd47-30b62f198c7c" containerName="horizon" Oct 07 23:52:20 crc kubenswrapper[4871]: E1007 23:52:20.409618 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.409626 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.409892 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.409913 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="75705323-d7c5-4134-bd47-30b62f198c7c" containerName="horizon" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.409931 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b251d66-e2c5-43ab-81db-339b0690d2dc" containerName="horizon-log" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.409944 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="75705323-d7c5-4134-bd47-30b62f198c7c" containerName="horizon-log" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.411296 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.431296 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f455d5855-8bgr8"] Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.585016 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-logs\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.585077 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-horizon-secret-key\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.585105 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgr6r\" (UniqueName: \"kubernetes.io/projected/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-kube-api-access-wgr6r\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.585140 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-config-data\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.585180 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-scripts\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.687256 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-config-data\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.687347 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-scripts\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.687550 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-logs\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.687584 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-horizon-secret-key\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.687608 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgr6r\" (UniqueName: \"kubernetes.io/projected/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-kube-api-access-wgr6r\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.688946 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-logs\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.688991 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-scripts\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.689038 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-config-data\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.697628 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-horizon-secret-key\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.707480 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgr6r\" (UniqueName: \"kubernetes.io/projected/674d0e65-ecd4-4b3a-a734-4c01e2b4192b-kube-api-access-wgr6r\") pod \"horizon-7f455d5855-8bgr8\" (UID: \"674d0e65-ecd4-4b3a-a734-4c01e2b4192b\") " pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:20 crc kubenswrapper[4871]: I1007 23:52:20.734476 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:21 crc kubenswrapper[4871]: I1007 23:52:21.033878 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-2gqc9"] Oct 07 23:52:21 crc kubenswrapper[4871]: I1007 23:52:21.041727 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-2gqc9"] Oct 07 23:52:21 crc kubenswrapper[4871]: I1007 23:52:21.191043 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f455d5855-8bgr8"] Oct 07 23:52:21 crc kubenswrapper[4871]: I1007 23:52:21.834604 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-pdkts"] Oct 07 23:52:21 crc kubenswrapper[4871]: I1007 23:52:21.836468 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pdkts" Oct 07 23:52:21 crc kubenswrapper[4871]: I1007 23:52:21.844643 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pdkts"] Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.018032 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn989\" (UniqueName: \"kubernetes.io/projected/ba09d7fe-5976-4a5b-8f7e-fbb0988e335f-kube-api-access-jn989\") pod \"heat-db-create-pdkts\" (UID: \"ba09d7fe-5976-4a5b-8f7e-fbb0988e335f\") " pod="openstack/heat-db-create-pdkts" Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.112638 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f455d5855-8bgr8" event={"ID":"674d0e65-ecd4-4b3a-a734-4c01e2b4192b","Type":"ContainerStarted","Data":"405de39b43cc118da26632c3f6601d3b5bab6a31802388048b832c27510f772b"} Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.112705 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f455d5855-8bgr8" event={"ID":"674d0e65-ecd4-4b3a-a734-4c01e2b4192b","Type":"ContainerStarted","Data":"5c2357c1df829463fca7dd0def9b8c0d84985c0227057ae975d1a55ad48a06f0"} Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.112721 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f455d5855-8bgr8" event={"ID":"674d0e65-ecd4-4b3a-a734-4c01e2b4192b","Type":"ContainerStarted","Data":"91846b57ec31ada3d1473563587f8cb3cd413793f58c5d3ced17b223b6e28bca"} Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.119986 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn989\" (UniqueName: \"kubernetes.io/projected/ba09d7fe-5976-4a5b-8f7e-fbb0988e335f-kube-api-access-jn989\") pod \"heat-db-create-pdkts\" (UID: \"ba09d7fe-5976-4a5b-8f7e-fbb0988e335f\") " pod="openstack/heat-db-create-pdkts" Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.142520 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn989\" (UniqueName: \"kubernetes.io/projected/ba09d7fe-5976-4a5b-8f7e-fbb0988e335f-kube-api-access-jn989\") pod \"heat-db-create-pdkts\" (UID: \"ba09d7fe-5976-4a5b-8f7e-fbb0988e335f\") " pod="openstack/heat-db-create-pdkts" Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.152479 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f455d5855-8bgr8" podStartSLOduration=2.152458736 podStartE2EDuration="2.152458736s" podCreationTimestamp="2025-10-07 23:52:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:52:22.142731911 +0000 UTC m=+6215.945430004" watchObservedRunningTime="2025-10-07 23:52:22.152458736 +0000 UTC m=+6215.955156819" Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.169284 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pdkts" Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.633495 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pdkts"] Oct 07 23:52:22 crc kubenswrapper[4871]: I1007 23:52:22.993309 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9b014be-6d59-4c4c-a932-3d5a9efe4b12" path="/var/lib/kubelet/pods/e9b014be-6d59-4c4c-a932-3d5a9efe4b12/volumes" Oct 07 23:52:23 crc kubenswrapper[4871]: I1007 23:52:23.122634 4871 generic.go:334] "Generic (PLEG): container finished" podID="ba09d7fe-5976-4a5b-8f7e-fbb0988e335f" containerID="a0584b9a7c43e6906ad7b1643ab46010101ba18fa1118e85ed57db2ad65ae6c5" exitCode=0 Oct 07 23:52:23 crc kubenswrapper[4871]: I1007 23:52:23.122673 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pdkts" event={"ID":"ba09d7fe-5976-4a5b-8f7e-fbb0988e335f","Type":"ContainerDied","Data":"a0584b9a7c43e6906ad7b1643ab46010101ba18fa1118e85ed57db2ad65ae6c5"} Oct 07 23:52:23 crc kubenswrapper[4871]: I1007 23:52:23.122974 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pdkts" event={"ID":"ba09d7fe-5976-4a5b-8f7e-fbb0988e335f","Type":"ContainerStarted","Data":"7c99328a2b54cb7d0cd6221a134466f20e38d7adf9b0547af89673bc6f8c2e07"} Oct 07 23:52:24 crc kubenswrapper[4871]: I1007 23:52:24.599725 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pdkts" Oct 07 23:52:24 crc kubenswrapper[4871]: I1007 23:52:24.770267 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn989\" (UniqueName: \"kubernetes.io/projected/ba09d7fe-5976-4a5b-8f7e-fbb0988e335f-kube-api-access-jn989\") pod \"ba09d7fe-5976-4a5b-8f7e-fbb0988e335f\" (UID: \"ba09d7fe-5976-4a5b-8f7e-fbb0988e335f\") " Oct 07 23:52:24 crc kubenswrapper[4871]: I1007 23:52:24.779855 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba09d7fe-5976-4a5b-8f7e-fbb0988e335f-kube-api-access-jn989" (OuterVolumeSpecName: "kube-api-access-jn989") pod "ba09d7fe-5976-4a5b-8f7e-fbb0988e335f" (UID: "ba09d7fe-5976-4a5b-8f7e-fbb0988e335f"). InnerVolumeSpecName "kube-api-access-jn989". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:52:24 crc kubenswrapper[4871]: I1007 23:52:24.872874 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn989\" (UniqueName: \"kubernetes.io/projected/ba09d7fe-5976-4a5b-8f7e-fbb0988e335f-kube-api-access-jn989\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:25 crc kubenswrapper[4871]: I1007 23:52:25.150083 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pdkts" event={"ID":"ba09d7fe-5976-4a5b-8f7e-fbb0988e335f","Type":"ContainerDied","Data":"7c99328a2b54cb7d0cd6221a134466f20e38d7adf9b0547af89673bc6f8c2e07"} Oct 07 23:52:25 crc kubenswrapper[4871]: I1007 23:52:25.150354 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c99328a2b54cb7d0cd6221a134466f20e38d7adf9b0547af89673bc6f8c2e07" Oct 07 23:52:25 crc kubenswrapper[4871]: I1007 23:52:25.150116 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pdkts" Oct 07 23:52:30 crc kubenswrapper[4871]: I1007 23:52:30.735504 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:30 crc kubenswrapper[4871]: I1007 23:52:30.736094 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:31 crc kubenswrapper[4871]: I1007 23:52:31.961817 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-a7bc-account-create-f4rjs"] Oct 07 23:52:31 crc kubenswrapper[4871]: E1007 23:52:31.962650 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba09d7fe-5976-4a5b-8f7e-fbb0988e335f" containerName="mariadb-database-create" Oct 07 23:52:31 crc kubenswrapper[4871]: I1007 23:52:31.962670 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba09d7fe-5976-4a5b-8f7e-fbb0988e335f" containerName="mariadb-database-create" Oct 07 23:52:31 crc kubenswrapper[4871]: I1007 23:52:31.963009 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba09d7fe-5976-4a5b-8f7e-fbb0988e335f" containerName="mariadb-database-create" Oct 07 23:52:31 crc kubenswrapper[4871]: I1007 23:52:31.963770 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a7bc-account-create-f4rjs" Oct 07 23:52:31 crc kubenswrapper[4871]: I1007 23:52:31.969708 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 07 23:52:31 crc kubenswrapper[4871]: I1007 23:52:31.980564 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a7bc-account-create-f4rjs"] Oct 07 23:52:32 crc kubenswrapper[4871]: I1007 23:52:32.134139 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5spsz\" (UniqueName: \"kubernetes.io/projected/5c55d7c7-45dd-4e70-b1fe-cf549af69d58-kube-api-access-5spsz\") pod \"heat-a7bc-account-create-f4rjs\" (UID: \"5c55d7c7-45dd-4e70-b1fe-cf549af69d58\") " pod="openstack/heat-a7bc-account-create-f4rjs" Oct 07 23:52:32 crc kubenswrapper[4871]: I1007 23:52:32.236336 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5spsz\" (UniqueName: \"kubernetes.io/projected/5c55d7c7-45dd-4e70-b1fe-cf549af69d58-kube-api-access-5spsz\") pod \"heat-a7bc-account-create-f4rjs\" (UID: \"5c55d7c7-45dd-4e70-b1fe-cf549af69d58\") " pod="openstack/heat-a7bc-account-create-f4rjs" Oct 07 23:52:32 crc kubenswrapper[4871]: I1007 23:52:32.270487 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5spsz\" (UniqueName: \"kubernetes.io/projected/5c55d7c7-45dd-4e70-b1fe-cf549af69d58-kube-api-access-5spsz\") pod \"heat-a7bc-account-create-f4rjs\" (UID: \"5c55d7c7-45dd-4e70-b1fe-cf549af69d58\") " pod="openstack/heat-a7bc-account-create-f4rjs" Oct 07 23:52:32 crc kubenswrapper[4871]: I1007 23:52:32.292970 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a7bc-account-create-f4rjs" Oct 07 23:52:32 crc kubenswrapper[4871]: I1007 23:52:32.805426 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a7bc-account-create-f4rjs"] Oct 07 23:52:33 crc kubenswrapper[4871]: I1007 23:52:33.243583 4871 generic.go:334] "Generic (PLEG): container finished" podID="5c55d7c7-45dd-4e70-b1fe-cf549af69d58" containerID="d9a8ad4187bab8ad6a9547b58dc52e9469af380516890ae99f8b40fc3a53f443" exitCode=0 Oct 07 23:52:33 crc kubenswrapper[4871]: I1007 23:52:33.243656 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a7bc-account-create-f4rjs" event={"ID":"5c55d7c7-45dd-4e70-b1fe-cf549af69d58","Type":"ContainerDied","Data":"d9a8ad4187bab8ad6a9547b58dc52e9469af380516890ae99f8b40fc3a53f443"} Oct 07 23:52:33 crc kubenswrapper[4871]: I1007 23:52:33.243697 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a7bc-account-create-f4rjs" event={"ID":"5c55d7c7-45dd-4e70-b1fe-cf549af69d58","Type":"ContainerStarted","Data":"d3e1e1768e4de618e0546493edffd3c7d5448bebad970f6a653671513cbc41c5"} Oct 07 23:52:34 crc kubenswrapper[4871]: I1007 23:52:34.723027 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a7bc-account-create-f4rjs" Oct 07 23:52:34 crc kubenswrapper[4871]: I1007 23:52:34.807254 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5spsz\" (UniqueName: \"kubernetes.io/projected/5c55d7c7-45dd-4e70-b1fe-cf549af69d58-kube-api-access-5spsz\") pod \"5c55d7c7-45dd-4e70-b1fe-cf549af69d58\" (UID: \"5c55d7c7-45dd-4e70-b1fe-cf549af69d58\") " Oct 07 23:52:34 crc kubenswrapper[4871]: I1007 23:52:34.813817 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c55d7c7-45dd-4e70-b1fe-cf549af69d58-kube-api-access-5spsz" (OuterVolumeSpecName: "kube-api-access-5spsz") pod "5c55d7c7-45dd-4e70-b1fe-cf549af69d58" (UID: "5c55d7c7-45dd-4e70-b1fe-cf549af69d58"). InnerVolumeSpecName "kube-api-access-5spsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:52:34 crc kubenswrapper[4871]: I1007 23:52:34.910451 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5spsz\" (UniqueName: \"kubernetes.io/projected/5c55d7c7-45dd-4e70-b1fe-cf549af69d58-kube-api-access-5spsz\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:35 crc kubenswrapper[4871]: I1007 23:52:35.273133 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a7bc-account-create-f4rjs" Oct 07 23:52:35 crc kubenswrapper[4871]: I1007 23:52:35.273077 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a7bc-account-create-f4rjs" event={"ID":"5c55d7c7-45dd-4e70-b1fe-cf549af69d58","Type":"ContainerDied","Data":"d3e1e1768e4de618e0546493edffd3c7d5448bebad970f6a653671513cbc41c5"} Oct 07 23:52:35 crc kubenswrapper[4871]: I1007 23:52:35.276163 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3e1e1768e4de618e0546493edffd3c7d5448bebad970f6a653671513cbc41c5" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.044960 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-f5qn6"] Oct 07 23:52:37 crc kubenswrapper[4871]: E1007 23:52:37.046044 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c55d7c7-45dd-4e70-b1fe-cf549af69d58" containerName="mariadb-account-create" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.046062 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c55d7c7-45dd-4e70-b1fe-cf549af69d58" containerName="mariadb-account-create" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.046306 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c55d7c7-45dd-4e70-b1fe-cf549af69d58" containerName="mariadb-account-create" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.047171 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.051067 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.051347 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-2fvrj" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.076334 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-f5qn6"] Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.164636 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-config-data\") pod \"heat-db-sync-f5qn6\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.164725 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-852v7\" (UniqueName: \"kubernetes.io/projected/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-kube-api-access-852v7\") pod \"heat-db-sync-f5qn6\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.164807 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-combined-ca-bundle\") pod \"heat-db-sync-f5qn6\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.266212 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-combined-ca-bundle\") pod \"heat-db-sync-f5qn6\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.266381 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-config-data\") pod \"heat-db-sync-f5qn6\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.266453 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-852v7\" (UniqueName: \"kubernetes.io/projected/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-kube-api-access-852v7\") pod \"heat-db-sync-f5qn6\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.272349 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-combined-ca-bundle\") pod \"heat-db-sync-f5qn6\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.280902 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-config-data\") pod \"heat-db-sync-f5qn6\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.305664 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-852v7\" (UniqueName: \"kubernetes.io/projected/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-kube-api-access-852v7\") pod \"heat-db-sync-f5qn6\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.387976 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:37 crc kubenswrapper[4871]: I1007 23:52:37.877619 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-f5qn6"] Oct 07 23:52:37 crc kubenswrapper[4871]: W1007 23:52:37.884806 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb72cc287_a1e9_4b19_a4b2_7ef096b6d324.slice/crio-1112f9e590800eb75fed389c05c77823d92754228d0408ed29b6b03803647708 WatchSource:0}: Error finding container 1112f9e590800eb75fed389c05c77823d92754228d0408ed29b6b03803647708: Status 404 returned error can't find the container with id 1112f9e590800eb75fed389c05c77823d92754228d0408ed29b6b03803647708 Oct 07 23:52:38 crc kubenswrapper[4871]: I1007 23:52:38.311877 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-f5qn6" event={"ID":"b72cc287-a1e9-4b19-a4b2-7ef096b6d324","Type":"ContainerStarted","Data":"1112f9e590800eb75fed389c05c77823d92754228d0408ed29b6b03803647708"} Oct 07 23:52:42 crc kubenswrapper[4871]: I1007 23:52:42.517873 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:43 crc kubenswrapper[4871]: I1007 23:52:43.963496 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7f455d5855-8bgr8" Oct 07 23:52:44 crc kubenswrapper[4871]: I1007 23:52:44.032555 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-668c994789-hrcgh"] Oct 07 23:52:44 crc kubenswrapper[4871]: I1007 23:52:44.036186 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-668c994789-hrcgh" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon-log" containerID="cri-o://a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f" gracePeriod=30 Oct 07 23:52:44 crc kubenswrapper[4871]: I1007 23:52:44.036521 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-668c994789-hrcgh" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon" containerID="cri-o://713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc" gracePeriod=30 Oct 07 23:52:44 crc kubenswrapper[4871]: I1007 23:52:44.378563 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-f5qn6" event={"ID":"b72cc287-a1e9-4b19-a4b2-7ef096b6d324","Type":"ContainerStarted","Data":"9a1263606d6d8da8935a1a8f8b6f784f7b769c8bd6f3b01b10d3b85f87e0b314"} Oct 07 23:52:44 crc kubenswrapper[4871]: I1007 23:52:44.398386 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-f5qn6" podStartSLOduration=1.369408115 podStartE2EDuration="7.398366829s" podCreationTimestamp="2025-10-07 23:52:37 +0000 UTC" firstStartedPulling="2025-10-07 23:52:37.887934955 +0000 UTC m=+6231.690633029" lastFinishedPulling="2025-10-07 23:52:43.91689363 +0000 UTC m=+6237.719591743" observedRunningTime="2025-10-07 23:52:44.395857763 +0000 UTC m=+6238.198555846" watchObservedRunningTime="2025-10-07 23:52:44.398366829 +0000 UTC m=+6238.201064902" Oct 07 23:52:46 crc kubenswrapper[4871]: I1007 23:52:46.398771 4871 generic.go:334] "Generic (PLEG): container finished" podID="b72cc287-a1e9-4b19-a4b2-7ef096b6d324" containerID="9a1263606d6d8da8935a1a8f8b6f784f7b769c8bd6f3b01b10d3b85f87e0b314" exitCode=0 Oct 07 23:52:46 crc kubenswrapper[4871]: I1007 23:52:46.398874 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-f5qn6" event={"ID":"b72cc287-a1e9-4b19-a4b2-7ef096b6d324","Type":"ContainerDied","Data":"9a1263606d6d8da8935a1a8f8b6f784f7b769c8bd6f3b01b10d3b85f87e0b314"} Oct 07 23:52:47 crc kubenswrapper[4871]: I1007 23:52:47.409555 4871 generic.go:334] "Generic (PLEG): container finished" podID="dca91af8-6003-4f54-ba25-d85809b9da33" containerID="713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc" exitCode=0 Oct 07 23:52:47 crc kubenswrapper[4871]: I1007 23:52:47.409653 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-668c994789-hrcgh" event={"ID":"dca91af8-6003-4f54-ba25-d85809b9da33","Type":"ContainerDied","Data":"713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc"} Oct 07 23:52:47 crc kubenswrapper[4871]: I1007 23:52:47.834311 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:47 crc kubenswrapper[4871]: I1007 23:52:47.919501 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-852v7\" (UniqueName: \"kubernetes.io/projected/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-kube-api-access-852v7\") pod \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " Oct 07 23:52:47 crc kubenswrapper[4871]: I1007 23:52:47.919645 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-combined-ca-bundle\") pod \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " Oct 07 23:52:47 crc kubenswrapper[4871]: I1007 23:52:47.919848 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-config-data\") pod \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\" (UID: \"b72cc287-a1e9-4b19-a4b2-7ef096b6d324\") " Oct 07 23:52:47 crc kubenswrapper[4871]: I1007 23:52:47.931165 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-kube-api-access-852v7" (OuterVolumeSpecName: "kube-api-access-852v7") pod "b72cc287-a1e9-4b19-a4b2-7ef096b6d324" (UID: "b72cc287-a1e9-4b19-a4b2-7ef096b6d324"). InnerVolumeSpecName "kube-api-access-852v7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:52:47 crc kubenswrapper[4871]: I1007 23:52:47.960102 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b72cc287-a1e9-4b19-a4b2-7ef096b6d324" (UID: "b72cc287-a1e9-4b19-a4b2-7ef096b6d324"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:52:48 crc kubenswrapper[4871]: I1007 23:52:48.015375 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-config-data" (OuterVolumeSpecName: "config-data") pod "b72cc287-a1e9-4b19-a4b2-7ef096b6d324" (UID: "b72cc287-a1e9-4b19-a4b2-7ef096b6d324"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:52:48 crc kubenswrapper[4871]: I1007 23:52:48.023329 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-852v7\" (UniqueName: \"kubernetes.io/projected/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-kube-api-access-852v7\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:48 crc kubenswrapper[4871]: I1007 23:52:48.023370 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:48 crc kubenswrapper[4871]: I1007 23:52:48.023390 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b72cc287-a1e9-4b19-a4b2-7ef096b6d324-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:52:48 crc kubenswrapper[4871]: I1007 23:52:48.422241 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-f5qn6" event={"ID":"b72cc287-a1e9-4b19-a4b2-7ef096b6d324","Type":"ContainerDied","Data":"1112f9e590800eb75fed389c05c77823d92754228d0408ed29b6b03803647708"} Oct 07 23:52:48 crc kubenswrapper[4871]: I1007 23:52:48.422284 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1112f9e590800eb75fed389c05c77823d92754228d0408ed29b6b03803647708" Oct 07 23:52:48 crc kubenswrapper[4871]: I1007 23:52:48.422358 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-f5qn6" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.782633 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-574ddff48-pw8c9"] Oct 07 23:52:49 crc kubenswrapper[4871]: E1007 23:52:49.783955 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72cc287-a1e9-4b19-a4b2-7ef096b6d324" containerName="heat-db-sync" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.783989 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72cc287-a1e9-4b19-a4b2-7ef096b6d324" containerName="heat-db-sync" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.784588 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b72cc287-a1e9-4b19-a4b2-7ef096b6d324" containerName="heat-db-sync" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.787125 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.790689 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.791054 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.792974 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-2fvrj" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.810726 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-574ddff48-pw8c9"] Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.870077 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-config-data-custom\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.870994 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-config-data\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.871908 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-combined-ca-bundle\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.881272 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wjsc\" (UniqueName: \"kubernetes.io/projected/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-kube-api-access-8wjsc\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.935124 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5c75875cf4-v2bhz"] Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.937258 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.942194 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.963022 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5c75875cf4-v2bhz"] Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.977850 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5b8c8579fc-wmrqt"] Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.979492 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.982019 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.983933 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-config-data\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.983997 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0bd4d03-7aba-460f-84a0-1fa48a81e822-config-data-custom\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.984205 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-combined-ca-bundle\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.984250 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0bd4d03-7aba-460f-84a0-1fa48a81e822-config-data\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.984293 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42v5g\" (UniqueName: \"kubernetes.io/projected/c0bd4d03-7aba-460f-84a0-1fa48a81e822-kube-api-access-42v5g\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.984378 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wjsc\" (UniqueName: \"kubernetes.io/projected/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-kube-api-access-8wjsc\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.984425 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-config-data-custom\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.984458 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0bd4d03-7aba-460f-84a0-1fa48a81e822-combined-ca-bundle\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.989859 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-combined-ca-bundle\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.990027 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-config-data\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:49 crc kubenswrapper[4871]: I1007 23:52:49.990416 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-config-data-custom\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.014228 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wjsc\" (UniqueName: \"kubernetes.io/projected/e92c3a0b-ecce-4129-9cb4-d6bf8f597225-kube-api-access-8wjsc\") pod \"heat-engine-574ddff48-pw8c9\" (UID: \"e92c3a0b-ecce-4129-9cb4-d6bf8f597225\") " pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.021634 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5b8c8579fc-wmrqt"] Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.085888 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzpdh\" (UniqueName: \"kubernetes.io/projected/0193e1f0-49ff-4945-a039-a31e3431dec3-kube-api-access-hzpdh\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.085933 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0193e1f0-49ff-4945-a039-a31e3431dec3-config-data\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.086099 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0bd4d03-7aba-460f-84a0-1fa48a81e822-combined-ca-bundle\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.086262 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0bd4d03-7aba-460f-84a0-1fa48a81e822-config-data-custom\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.086510 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0193e1f0-49ff-4945-a039-a31e3431dec3-combined-ca-bundle\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.086667 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0193e1f0-49ff-4945-a039-a31e3431dec3-config-data-custom\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.086862 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0bd4d03-7aba-460f-84a0-1fa48a81e822-config-data\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.086948 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42v5g\" (UniqueName: \"kubernetes.io/projected/c0bd4d03-7aba-460f-84a0-1fa48a81e822-kube-api-access-42v5g\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.090898 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0bd4d03-7aba-460f-84a0-1fa48a81e822-combined-ca-bundle\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.092607 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0bd4d03-7aba-460f-84a0-1fa48a81e822-config-data-custom\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.095106 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0bd4d03-7aba-460f-84a0-1fa48a81e822-config-data\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.107629 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.107929 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42v5g\" (UniqueName: \"kubernetes.io/projected/c0bd4d03-7aba-460f-84a0-1fa48a81e822-kube-api-access-42v5g\") pod \"heat-cfnapi-5c75875cf4-v2bhz\" (UID: \"c0bd4d03-7aba-460f-84a0-1fa48a81e822\") " pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.193044 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzpdh\" (UniqueName: \"kubernetes.io/projected/0193e1f0-49ff-4945-a039-a31e3431dec3-kube-api-access-hzpdh\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.193364 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0193e1f0-49ff-4945-a039-a31e3431dec3-config-data\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.193855 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0193e1f0-49ff-4945-a039-a31e3431dec3-combined-ca-bundle\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.193928 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0193e1f0-49ff-4945-a039-a31e3431dec3-config-data-custom\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.200727 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0193e1f0-49ff-4945-a039-a31e3431dec3-config-data-custom\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.206853 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0193e1f0-49ff-4945-a039-a31e3431dec3-combined-ca-bundle\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.208126 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0193e1f0-49ff-4945-a039-a31e3431dec3-config-data\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.212457 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzpdh\" (UniqueName: \"kubernetes.io/projected/0193e1f0-49ff-4945-a039-a31e3431dec3-kube-api-access-hzpdh\") pod \"heat-api-5b8c8579fc-wmrqt\" (UID: \"0193e1f0-49ff-4945-a039-a31e3431dec3\") " pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.260552 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.389778 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.640406 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-574ddff48-pw8c9"] Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.790648 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5c75875cf4-v2bhz"] Oct 07 23:52:50 crc kubenswrapper[4871]: W1007 23:52:50.802363 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0bd4d03_7aba_460f_84a0_1fa48a81e822.slice/crio-7cc399db6c1c001a367cfe299491582f8a6860f172646c12c8c568a4d5427616 WatchSource:0}: Error finding container 7cc399db6c1c001a367cfe299491582f8a6860f172646c12c8c568a4d5427616: Status 404 returned error can't find the container with id 7cc399db6c1c001a367cfe299491582f8a6860f172646c12c8c568a4d5427616 Oct 07 23:52:50 crc kubenswrapper[4871]: I1007 23:52:50.913655 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5b8c8579fc-wmrqt"] Oct 07 23:52:51 crc kubenswrapper[4871]: I1007 23:52:51.454388 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5b8c8579fc-wmrqt" event={"ID":"0193e1f0-49ff-4945-a039-a31e3431dec3","Type":"ContainerStarted","Data":"7baa6a6d9a96d73ccef131997564c1e5b398601c1a70d21555a8235e11b877df"} Oct 07 23:52:51 crc kubenswrapper[4871]: I1007 23:52:51.457050 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-574ddff48-pw8c9" event={"ID":"e92c3a0b-ecce-4129-9cb4-d6bf8f597225","Type":"ContainerStarted","Data":"b3ac9f512168147b17fc5e587252e1d2388d7c09f9f2860e90fced6723b8f9ed"} Oct 07 23:52:51 crc kubenswrapper[4871]: I1007 23:52:51.457117 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-574ddff48-pw8c9" event={"ID":"e92c3a0b-ecce-4129-9cb4-d6bf8f597225","Type":"ContainerStarted","Data":"c1859a1fd57277c5460aa949fae9fe8df751dea51d219eb0fa80c84898853d42"} Oct 07 23:52:51 crc kubenswrapper[4871]: I1007 23:52:51.457162 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:52:51 crc kubenswrapper[4871]: I1007 23:52:51.458781 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" event={"ID":"c0bd4d03-7aba-460f-84a0-1fa48a81e822","Type":"ContainerStarted","Data":"7cc399db6c1c001a367cfe299491582f8a6860f172646c12c8c568a4d5427616"} Oct 07 23:52:51 crc kubenswrapper[4871]: I1007 23:52:51.481194 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-574ddff48-pw8c9" podStartSLOduration=2.481171075 podStartE2EDuration="2.481171075s" podCreationTimestamp="2025-10-07 23:52:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:52:51.474154181 +0000 UTC m=+6245.276852244" watchObservedRunningTime="2025-10-07 23:52:51.481171075 +0000 UTC m=+6245.283869148" Oct 07 23:52:52 crc kubenswrapper[4871]: I1007 23:52:52.834157 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-668c994789-hrcgh" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Oct 07 23:52:53 crc kubenswrapper[4871]: I1007 23:52:53.494405 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5b8c8579fc-wmrqt" event={"ID":"0193e1f0-49ff-4945-a039-a31e3431dec3","Type":"ContainerStarted","Data":"f6c1c92ea2be755dad4601b3b183c63c788b3152ff7bcb1423e1c6e34a1357e7"} Oct 07 23:52:53 crc kubenswrapper[4871]: I1007 23:52:53.499810 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" event={"ID":"c0bd4d03-7aba-460f-84a0-1fa48a81e822","Type":"ContainerStarted","Data":"216843f478bcca0c4b978602b67b56e45e3cfd9fff48f24e9384f3a323beb71b"} Oct 07 23:52:53 crc kubenswrapper[4871]: I1007 23:52:53.499894 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:52:53 crc kubenswrapper[4871]: I1007 23:52:53.532190 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5b8c8579fc-wmrqt" podStartSLOduration=2.481033212 podStartE2EDuration="4.532171461s" podCreationTimestamp="2025-10-07 23:52:49 +0000 UTC" firstStartedPulling="2025-10-07 23:52:50.945293469 +0000 UTC m=+6244.747991542" lastFinishedPulling="2025-10-07 23:52:52.996431718 +0000 UTC m=+6246.799129791" observedRunningTime="2025-10-07 23:52:53.509725282 +0000 UTC m=+6247.312423355" watchObservedRunningTime="2025-10-07 23:52:53.532171461 +0000 UTC m=+6247.334869534" Oct 07 23:52:53 crc kubenswrapper[4871]: I1007 23:52:53.556647 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" podStartSLOduration=2.368475119 podStartE2EDuration="4.556624002s" podCreationTimestamp="2025-10-07 23:52:49 +0000 UTC" firstStartedPulling="2025-10-07 23:52:50.805000199 +0000 UTC m=+6244.607698272" lastFinishedPulling="2025-10-07 23:52:52.993149082 +0000 UTC m=+6246.795847155" observedRunningTime="2025-10-07 23:52:53.529582403 +0000 UTC m=+6247.332280476" watchObservedRunningTime="2025-10-07 23:52:53.556624002 +0000 UTC m=+6247.359322075" Oct 07 23:52:54 crc kubenswrapper[4871]: I1007 23:52:54.509844 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:52:59 crc kubenswrapper[4871]: I1007 23:52:59.357749 4871 scope.go:117] "RemoveContainer" containerID="c933ceb6c23a048767fe87f97d851a5718a6bac3cda3abfa6cbd7c3bff7a0d87" Oct 07 23:52:59 crc kubenswrapper[4871]: I1007 23:52:59.384197 4871 scope.go:117] "RemoveContainer" containerID="2bca1fdad9767568e53e746fac1cec950cf85da925f66f971a0e074a54250a6a" Oct 07 23:52:59 crc kubenswrapper[4871]: I1007 23:52:59.455748 4871 scope.go:117] "RemoveContainer" containerID="4b3968107feab7d9fc685588f70fb5321e0c37bb16c7db451fb28831edc7fa53" Oct 07 23:53:00 crc kubenswrapper[4871]: I1007 23:53:00.149879 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-574ddff48-pw8c9" Oct 07 23:53:01 crc kubenswrapper[4871]: I1007 23:53:01.568133 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5c75875cf4-v2bhz" Oct 07 23:53:01 crc kubenswrapper[4871]: I1007 23:53:01.632705 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5b8c8579fc-wmrqt" Oct 07 23:53:02 crc kubenswrapper[4871]: I1007 23:53:02.834360 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-668c994789-hrcgh" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Oct 07 23:53:12 crc kubenswrapper[4871]: I1007 23:53:12.834574 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-668c994789-hrcgh" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Oct 07 23:53:12 crc kubenswrapper[4871]: I1007 23:53:12.835714 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.542569 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.729435 4871 generic.go:334] "Generic (PLEG): container finished" podID="dca91af8-6003-4f54-ba25-d85809b9da33" containerID="a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f" exitCode=137 Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.729495 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-668c994789-hrcgh" event={"ID":"dca91af8-6003-4f54-ba25-d85809b9da33","Type":"ContainerDied","Data":"a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f"} Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.729523 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-668c994789-hrcgh" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.729541 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-668c994789-hrcgh" event={"ID":"dca91af8-6003-4f54-ba25-d85809b9da33","Type":"ContainerDied","Data":"be8d11dca13a4ce476cdf66892063681825c5fca331165ce9ade3a2729f3c2fa"} Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.729574 4871 scope.go:117] "RemoveContainer" containerID="713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.741377 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dca91af8-6003-4f54-ba25-d85809b9da33-horizon-secret-key\") pod \"dca91af8-6003-4f54-ba25-d85809b9da33\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.741418 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dca91af8-6003-4f54-ba25-d85809b9da33-logs\") pod \"dca91af8-6003-4f54-ba25-d85809b9da33\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.741464 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kn42\" (UniqueName: \"kubernetes.io/projected/dca91af8-6003-4f54-ba25-d85809b9da33-kube-api-access-6kn42\") pod \"dca91af8-6003-4f54-ba25-d85809b9da33\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.741606 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-config-data\") pod \"dca91af8-6003-4f54-ba25-d85809b9da33\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.741686 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-scripts\") pod \"dca91af8-6003-4f54-ba25-d85809b9da33\" (UID: \"dca91af8-6003-4f54-ba25-d85809b9da33\") " Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.741961 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dca91af8-6003-4f54-ba25-d85809b9da33-logs" (OuterVolumeSpecName: "logs") pod "dca91af8-6003-4f54-ba25-d85809b9da33" (UID: "dca91af8-6003-4f54-ba25-d85809b9da33"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.742268 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dca91af8-6003-4f54-ba25-d85809b9da33-logs\") on node \"crc\" DevicePath \"\"" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.747057 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dca91af8-6003-4f54-ba25-d85809b9da33-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "dca91af8-6003-4f54-ba25-d85809b9da33" (UID: "dca91af8-6003-4f54-ba25-d85809b9da33"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.748591 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dca91af8-6003-4f54-ba25-d85809b9da33-kube-api-access-6kn42" (OuterVolumeSpecName: "kube-api-access-6kn42") pod "dca91af8-6003-4f54-ba25-d85809b9da33" (UID: "dca91af8-6003-4f54-ba25-d85809b9da33"). InnerVolumeSpecName "kube-api-access-6kn42". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.768807 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-config-data" (OuterVolumeSpecName: "config-data") pod "dca91af8-6003-4f54-ba25-d85809b9da33" (UID: "dca91af8-6003-4f54-ba25-d85809b9da33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.772216 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-scripts" (OuterVolumeSpecName: "scripts") pod "dca91af8-6003-4f54-ba25-d85809b9da33" (UID: "dca91af8-6003-4f54-ba25-d85809b9da33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.844257 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.844295 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dca91af8-6003-4f54-ba25-d85809b9da33-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.844312 4871 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dca91af8-6003-4f54-ba25-d85809b9da33-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.844327 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kn42\" (UniqueName: \"kubernetes.io/projected/dca91af8-6003-4f54-ba25-d85809b9da33-kube-api-access-6kn42\") on node \"crc\" DevicePath \"\"" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.880224 4871 scope.go:117] "RemoveContainer" containerID="a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.915695 4871 scope.go:117] "RemoveContainer" containerID="713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc" Oct 07 23:53:14 crc kubenswrapper[4871]: E1007 23:53:14.916231 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc\": container with ID starting with 713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc not found: ID does not exist" containerID="713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.916307 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc"} err="failed to get container status \"713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc\": rpc error: code = NotFound desc = could not find container \"713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc\": container with ID starting with 713acd8115e23b426e718e9ea633587cc1e37dd7ce0af0962c0f5b3176de72fc not found: ID does not exist" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.916353 4871 scope.go:117] "RemoveContainer" containerID="a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f" Oct 07 23:53:14 crc kubenswrapper[4871]: E1007 23:53:14.916961 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f\": container with ID starting with a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f not found: ID does not exist" containerID="a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f" Oct 07 23:53:14 crc kubenswrapper[4871]: I1007 23:53:14.917006 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f"} err="failed to get container status \"a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f\": rpc error: code = NotFound desc = could not find container \"a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f\": container with ID starting with a9c27e3f84c0afd2b56a4650753abf11d4c80e62da744ef8ccba29e72990ca9f not found: ID does not exist" Oct 07 23:53:15 crc kubenswrapper[4871]: I1007 23:53:15.060894 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-668c994789-hrcgh"] Oct 07 23:53:15 crc kubenswrapper[4871]: I1007 23:53:15.070029 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-668c994789-hrcgh"] Oct 07 23:53:16 crc kubenswrapper[4871]: I1007 23:53:16.997042 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" path="/var/lib/kubelet/pods/dca91af8-6003-4f54-ba25-d85809b9da33/volumes" Oct 07 23:53:17 crc kubenswrapper[4871]: I1007 23:53:17.056197 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-4hfxw"] Oct 07 23:53:17 crc kubenswrapper[4871]: I1007 23:53:17.064064 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-pdvjn"] Oct 07 23:53:17 crc kubenswrapper[4871]: I1007 23:53:17.070807 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-4hfxw"] Oct 07 23:53:17 crc kubenswrapper[4871]: I1007 23:53:17.079712 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-pdvjn"] Oct 07 23:53:18 crc kubenswrapper[4871]: I1007 23:53:18.027855 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-q6fp5"] Oct 07 23:53:18 crc kubenswrapper[4871]: I1007 23:53:18.036049 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-q6fp5"] Oct 07 23:53:18 crc kubenswrapper[4871]: I1007 23:53:18.994414 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49092041-b1e7-422f-9e09-b398fb63a548" path="/var/lib/kubelet/pods/49092041-b1e7-422f-9e09-b398fb63a548/volumes" Oct 07 23:53:18 crc kubenswrapper[4871]: I1007 23:53:18.995326 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5281c6fc-e54b-4eba-b710-b62ea4e2a9de" path="/var/lib/kubelet/pods/5281c6fc-e54b-4eba-b710-b62ea4e2a9de/volumes" Oct 07 23:53:18 crc kubenswrapper[4871]: I1007 23:53:18.996276 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7141cb8-92b3-4b46-b0f7-29fb0b48d479" path="/var/lib/kubelet/pods/b7141cb8-92b3-4b46-b0f7-29fb0b48d479/volumes" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.783963 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn"] Oct 07 23:53:24 crc kubenswrapper[4871]: E1007 23:53:24.785102 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.785118 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon" Oct 07 23:53:24 crc kubenswrapper[4871]: E1007 23:53:24.785172 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon-log" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.785199 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon-log" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.785442 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon-log" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.785473 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="dca91af8-6003-4f54-ba25-d85809b9da33" containerName="horizon" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.787228 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.789423 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.802427 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn"] Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.958396 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.958675 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:24 crc kubenswrapper[4871]: I1007 23:53:24.958860 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwlbh\" (UniqueName: \"kubernetes.io/projected/0ee6dc03-a9cc-468f-aea1-0ec712237561-kube-api-access-mwlbh\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:25 crc kubenswrapper[4871]: I1007 23:53:25.061202 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:25 crc kubenswrapper[4871]: I1007 23:53:25.061320 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwlbh\" (UniqueName: \"kubernetes.io/projected/0ee6dc03-a9cc-468f-aea1-0ec712237561-kube-api-access-mwlbh\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:25 crc kubenswrapper[4871]: I1007 23:53:25.061415 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:25 crc kubenswrapper[4871]: I1007 23:53:25.061852 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:25 crc kubenswrapper[4871]: I1007 23:53:25.061935 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:25 crc kubenswrapper[4871]: I1007 23:53:25.084679 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwlbh\" (UniqueName: \"kubernetes.io/projected/0ee6dc03-a9cc-468f-aea1-0ec712237561-kube-api-access-mwlbh\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:25 crc kubenswrapper[4871]: I1007 23:53:25.139218 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:25 crc kubenswrapper[4871]: I1007 23:53:25.634011 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn"] Oct 07 23:53:25 crc kubenswrapper[4871]: I1007 23:53:25.856004 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" event={"ID":"0ee6dc03-a9cc-468f-aea1-0ec712237561","Type":"ContainerStarted","Data":"50c9469a01fc9365c920df04b7192309aae67afc66b1a7f04d24f53baecf48b4"} Oct 07 23:53:26 crc kubenswrapper[4871]: E1007 23:53:26.507488 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ee6dc03_a9cc_468f_aea1_0ec712237561.slice/crio-conmon-e0b31e13ecf8644ecb2c30f657c1e869f159528bf7cdea30ad2594a045bffc22.scope\": RecentStats: unable to find data in memory cache]" Oct 07 23:53:26 crc kubenswrapper[4871]: I1007 23:53:26.869960 4871 generic.go:334] "Generic (PLEG): container finished" podID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerID="e0b31e13ecf8644ecb2c30f657c1e869f159528bf7cdea30ad2594a045bffc22" exitCode=0 Oct 07 23:53:26 crc kubenswrapper[4871]: I1007 23:53:26.870076 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" event={"ID":"0ee6dc03-a9cc-468f-aea1-0ec712237561","Type":"ContainerDied","Data":"e0b31e13ecf8644ecb2c30f657c1e869f159528bf7cdea30ad2594a045bffc22"} Oct 07 23:53:27 crc kubenswrapper[4871]: I1007 23:53:27.037838 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-7720-account-create-78zkj"] Oct 07 23:53:27 crc kubenswrapper[4871]: I1007 23:53:27.044754 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-7720-account-create-78zkj"] Oct 07 23:53:28 crc kubenswrapper[4871]: I1007 23:53:28.060706 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-73a7-account-create-j6h4l"] Oct 07 23:53:28 crc kubenswrapper[4871]: I1007 23:53:28.071335 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-77da-account-create-wr7bc"] Oct 07 23:53:28 crc kubenswrapper[4871]: I1007 23:53:28.079559 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-73a7-account-create-j6h4l"] Oct 07 23:53:28 crc kubenswrapper[4871]: I1007 23:53:28.086559 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-77da-account-create-wr7bc"] Oct 07 23:53:28 crc kubenswrapper[4871]: I1007 23:53:28.896894 4871 generic.go:334] "Generic (PLEG): container finished" podID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerID="abda1ba24bb65c65f505ea194dff4a27d31817b6f2cd29aea4fc7911ea7b5551" exitCode=0 Oct 07 23:53:28 crc kubenswrapper[4871]: I1007 23:53:28.896998 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" event={"ID":"0ee6dc03-a9cc-468f-aea1-0ec712237561","Type":"ContainerDied","Data":"abda1ba24bb65c65f505ea194dff4a27d31817b6f2cd29aea4fc7911ea7b5551"} Oct 07 23:53:28 crc kubenswrapper[4871]: I1007 23:53:28.998402 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26045870-04fb-4552-881d-eeb30715d747" path="/var/lib/kubelet/pods/26045870-04fb-4552-881d-eeb30715d747/volumes" Oct 07 23:53:29 crc kubenswrapper[4871]: I1007 23:53:29.001433 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="833593af-531f-4b85-ab9f-56d6a657de34" path="/var/lib/kubelet/pods/833593af-531f-4b85-ab9f-56d6a657de34/volumes" Oct 07 23:53:29 crc kubenswrapper[4871]: I1007 23:53:29.002753 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2ab3195-0fe3-46cf-b07b-ae48b41273e5" path="/var/lib/kubelet/pods/f2ab3195-0fe3-46cf-b07b-ae48b41273e5/volumes" Oct 07 23:53:29 crc kubenswrapper[4871]: I1007 23:53:29.917581 4871 generic.go:334] "Generic (PLEG): container finished" podID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerID="ab15f662621c83d22874cdaa385a3e2d858a9521ef6bebcd755a91ff825a1dd5" exitCode=0 Oct 07 23:53:29 crc kubenswrapper[4871]: I1007 23:53:29.917741 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" event={"ID":"0ee6dc03-a9cc-468f-aea1-0ec712237561","Type":"ContainerDied","Data":"ab15f662621c83d22874cdaa385a3e2d858a9521ef6bebcd755a91ff825a1dd5"} Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.337638 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.498895 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwlbh\" (UniqueName: \"kubernetes.io/projected/0ee6dc03-a9cc-468f-aea1-0ec712237561-kube-api-access-mwlbh\") pod \"0ee6dc03-a9cc-468f-aea1-0ec712237561\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.499326 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-bundle\") pod \"0ee6dc03-a9cc-468f-aea1-0ec712237561\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.499527 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-util\") pod \"0ee6dc03-a9cc-468f-aea1-0ec712237561\" (UID: \"0ee6dc03-a9cc-468f-aea1-0ec712237561\") " Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.501535 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-bundle" (OuterVolumeSpecName: "bundle") pod "0ee6dc03-a9cc-468f-aea1-0ec712237561" (UID: "0ee6dc03-a9cc-468f-aea1-0ec712237561"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.505329 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ee6dc03-a9cc-468f-aea1-0ec712237561-kube-api-access-mwlbh" (OuterVolumeSpecName: "kube-api-access-mwlbh") pod "0ee6dc03-a9cc-468f-aea1-0ec712237561" (UID: "0ee6dc03-a9cc-468f-aea1-0ec712237561"). InnerVolumeSpecName "kube-api-access-mwlbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.512298 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-util" (OuterVolumeSpecName: "util") pod "0ee6dc03-a9cc-468f-aea1-0ec712237561" (UID: "0ee6dc03-a9cc-468f-aea1-0ec712237561"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.602000 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwlbh\" (UniqueName: \"kubernetes.io/projected/0ee6dc03-a9cc-468f-aea1-0ec712237561-kube-api-access-mwlbh\") on node \"crc\" DevicePath \"\"" Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.602196 4871 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.602252 4871 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0ee6dc03-a9cc-468f-aea1-0ec712237561-util\") on node \"crc\" DevicePath \"\"" Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.945105 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" event={"ID":"0ee6dc03-a9cc-468f-aea1-0ec712237561","Type":"ContainerDied","Data":"50c9469a01fc9365c920df04b7192309aae67afc66b1a7f04d24f53baecf48b4"} Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.945479 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50c9469a01fc9365c920df04b7192309aae67afc66b1a7f04d24f53baecf48b4" Oct 07 23:53:31 crc kubenswrapper[4871]: I1007 23:53:31.945307 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn" Oct 07 23:53:38 crc kubenswrapper[4871]: I1007 23:53:38.049231 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cn4jv"] Oct 07 23:53:38 crc kubenswrapper[4871]: I1007 23:53:38.061591 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cn4jv"] Oct 07 23:53:38 crc kubenswrapper[4871]: I1007 23:53:38.997522 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b06442f4-d46a-4ef2-a6d8-14855ec660c6" path="/var/lib/kubelet/pods/b06442f4-d46a-4ef2-a6d8-14855ec660c6/volumes" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.252536 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-nr674"] Oct 07 23:53:41 crc kubenswrapper[4871]: E1007 23:53:41.253992 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerName="extract" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.254022 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerName="extract" Oct 07 23:53:41 crc kubenswrapper[4871]: E1007 23:53:41.254077 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerName="pull" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.254086 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerName="pull" Oct 07 23:53:41 crc kubenswrapper[4871]: E1007 23:53:41.254130 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerName="util" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.254140 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerName="util" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.254684 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee6dc03-a9cc-468f-aea1-0ec712237561" containerName="extract" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.255983 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-nr674" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.261014 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-l776v" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.261289 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.261481 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.321613 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-nr674"] Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.383377 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m"] Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.384923 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.397842 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-5n7kf" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.398964 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.409318 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c"] Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.410946 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.425803 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nptc9\" (UniqueName: \"kubernetes.io/projected/a3adbb28-7bd8-418c-94e4-e0a1ff7c8726-kube-api-access-nptc9\") pod \"obo-prometheus-operator-7c8cf85677-nr674\" (UID: \"a3adbb28-7bd8-418c-94e4-e0a1ff7c8726\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-nr674" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.435153 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m"] Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.454868 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c"] Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.527704 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nptc9\" (UniqueName: \"kubernetes.io/projected/a3adbb28-7bd8-418c-94e4-e0a1ff7c8726-kube-api-access-nptc9\") pod \"obo-prometheus-operator-7c8cf85677-nr674\" (UID: \"a3adbb28-7bd8-418c-94e4-e0a1ff7c8726\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-nr674" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.527769 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/caf90bb9-05c5-40da-be2d-b1c57ef8075f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-hrq9c\" (UID: \"caf90bb9-05c5-40da-be2d-b1c57ef8075f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.527827 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/caf90bb9-05c5-40da-be2d-b1c57ef8075f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-hrq9c\" (UID: \"caf90bb9-05c5-40da-be2d-b1c57ef8075f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.527886 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d3e2b5e-61e3-45ec-b39b-ec6aee52c207-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-ggr6m\" (UID: \"1d3e2b5e-61e3-45ec-b39b-ec6aee52c207\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.527933 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d3e2b5e-61e3-45ec-b39b-ec6aee52c207-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-ggr6m\" (UID: \"1d3e2b5e-61e3-45ec-b39b-ec6aee52c207\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.546876 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nptc9\" (UniqueName: \"kubernetes.io/projected/a3adbb28-7bd8-418c-94e4-e0a1ff7c8726-kube-api-access-nptc9\") pod \"obo-prometheus-operator-7c8cf85677-nr674\" (UID: \"a3adbb28-7bd8-418c-94e4-e0a1ff7c8726\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-nr674" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.604604 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-ckv6s"] Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.606083 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.607837 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-jv526" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.608012 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.618088 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-ckv6s"] Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.626851 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-nr674" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.630386 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/caf90bb9-05c5-40da-be2d-b1c57ef8075f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-hrq9c\" (UID: \"caf90bb9-05c5-40da-be2d-b1c57ef8075f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.630462 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/caf90bb9-05c5-40da-be2d-b1c57ef8075f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-hrq9c\" (UID: \"caf90bb9-05c5-40da-be2d-b1c57ef8075f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.630532 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d3e2b5e-61e3-45ec-b39b-ec6aee52c207-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-ggr6m\" (UID: \"1d3e2b5e-61e3-45ec-b39b-ec6aee52c207\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.630584 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d3e2b5e-61e3-45ec-b39b-ec6aee52c207-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-ggr6m\" (UID: \"1d3e2b5e-61e3-45ec-b39b-ec6aee52c207\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.639367 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/caf90bb9-05c5-40da-be2d-b1c57ef8075f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-hrq9c\" (UID: \"caf90bb9-05c5-40da-be2d-b1c57ef8075f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.643142 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d3e2b5e-61e3-45ec-b39b-ec6aee52c207-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-ggr6m\" (UID: \"1d3e2b5e-61e3-45ec-b39b-ec6aee52c207\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.644210 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/caf90bb9-05c5-40da-be2d-b1c57ef8075f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-hrq9c\" (UID: \"caf90bb9-05c5-40da-be2d-b1c57ef8075f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.647348 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d3e2b5e-61e3-45ec-b39b-ec6aee52c207-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8497998f8-ggr6m\" (UID: \"1d3e2b5e-61e3-45ec-b39b-ec6aee52c207\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.708264 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.736301 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cvmk\" (UniqueName: \"kubernetes.io/projected/07703b22-45ab-4158-9afe-a11f5f0f1e41-kube-api-access-2cvmk\") pod \"observability-operator-cc5f78dfc-ckv6s\" (UID: \"07703b22-45ab-4158-9afe-a11f5f0f1e41\") " pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.736374 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.736741 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/07703b22-45ab-4158-9afe-a11f5f0f1e41-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-ckv6s\" (UID: \"07703b22-45ab-4158-9afe-a11f5f0f1e41\") " pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.782752 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-cjmdr"] Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.784202 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.787036 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-fb4k6" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.798914 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-cjmdr"] Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.839178 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cvmk\" (UniqueName: \"kubernetes.io/projected/07703b22-45ab-4158-9afe-a11f5f0f1e41-kube-api-access-2cvmk\") pod \"observability-operator-cc5f78dfc-ckv6s\" (UID: \"07703b22-45ab-4158-9afe-a11f5f0f1e41\") " pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.839251 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/07703b22-45ab-4158-9afe-a11f5f0f1e41-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-ckv6s\" (UID: \"07703b22-45ab-4158-9afe-a11f5f0f1e41\") " pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.843408 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/07703b22-45ab-4158-9afe-a11f5f0f1e41-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-ckv6s\" (UID: \"07703b22-45ab-4158-9afe-a11f5f0f1e41\") " pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.854785 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cvmk\" (UniqueName: \"kubernetes.io/projected/07703b22-45ab-4158-9afe-a11f5f0f1e41-kube-api-access-2cvmk\") pod \"observability-operator-cc5f78dfc-ckv6s\" (UID: \"07703b22-45ab-4158-9afe-a11f5f0f1e41\") " pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.933032 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.941284 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/48ff787b-84fa-4f90-9d13-e853c115d1c8-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-cjmdr\" (UID: \"48ff787b-84fa-4f90-9d13-e853c115d1c8\") " pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:53:41 crc kubenswrapper[4871]: I1007 23:53:41.941664 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsh5n\" (UniqueName: \"kubernetes.io/projected/48ff787b-84fa-4f90-9d13-e853c115d1c8-kube-api-access-bsh5n\") pod \"perses-operator-54bc95c9fb-cjmdr\" (UID: \"48ff787b-84fa-4f90-9d13-e853c115d1c8\") " pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.044967 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsh5n\" (UniqueName: \"kubernetes.io/projected/48ff787b-84fa-4f90-9d13-e853c115d1c8-kube-api-access-bsh5n\") pod \"perses-operator-54bc95c9fb-cjmdr\" (UID: \"48ff787b-84fa-4f90-9d13-e853c115d1c8\") " pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.045298 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/48ff787b-84fa-4f90-9d13-e853c115d1c8-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-cjmdr\" (UID: \"48ff787b-84fa-4f90-9d13-e853c115d1c8\") " pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.046161 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/48ff787b-84fa-4f90-9d13-e853c115d1c8-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-cjmdr\" (UID: \"48ff787b-84fa-4f90-9d13-e853c115d1c8\") " pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.088735 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsh5n\" (UniqueName: \"kubernetes.io/projected/48ff787b-84fa-4f90-9d13-e853c115d1c8-kube-api-access-bsh5n\") pod \"perses-operator-54bc95c9fb-cjmdr\" (UID: \"48ff787b-84fa-4f90-9d13-e853c115d1c8\") " pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.172910 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.461190 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c"] Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.472018 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m"] Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.645966 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-nr674"] Oct 07 23:53:42 crc kubenswrapper[4871]: W1007 23:53:42.655708 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3adbb28_7bd8_418c_94e4_e0a1ff7c8726.slice/crio-641ec5072a0d2b76de568cb990e166e116eb053256f562bead7e81e500da26f5 WatchSource:0}: Error finding container 641ec5072a0d2b76de568cb990e166e116eb053256f562bead7e81e500da26f5: Status 404 returned error can't find the container with id 641ec5072a0d2b76de568cb990e166e116eb053256f562bead7e81e500da26f5 Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.752006 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-ckv6s"] Oct 07 23:53:42 crc kubenswrapper[4871]: I1007 23:53:42.762154 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-cjmdr"] Oct 07 23:53:43 crc kubenswrapper[4871]: I1007 23:53:43.077846 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-nr674" event={"ID":"a3adbb28-7bd8-418c-94e4-e0a1ff7c8726","Type":"ContainerStarted","Data":"641ec5072a0d2b76de568cb990e166e116eb053256f562bead7e81e500da26f5"} Oct 07 23:53:43 crc kubenswrapper[4871]: I1007 23:53:43.079416 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" event={"ID":"caf90bb9-05c5-40da-be2d-b1c57ef8075f","Type":"ContainerStarted","Data":"af4888d209b4c1e45a969789a2450af87c9e5602ee51b3572fef230a38edea16"} Oct 07 23:53:43 crc kubenswrapper[4871]: I1007 23:53:43.080771 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" event={"ID":"1d3e2b5e-61e3-45ec-b39b-ec6aee52c207","Type":"ContainerStarted","Data":"35efbaf9e651f39603041cfabb2dc2ea8dffaec86c85e9aa74dee13c2567092c"} Oct 07 23:53:43 crc kubenswrapper[4871]: I1007 23:53:43.082214 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" event={"ID":"07703b22-45ab-4158-9afe-a11f5f0f1e41","Type":"ContainerStarted","Data":"011ce69867693ed148251c0291262f8ae77b91df010a9f9c01bb14e2e20a8d35"} Oct 07 23:53:43 crc kubenswrapper[4871]: I1007 23:53:43.083569 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" event={"ID":"48ff787b-84fa-4f90-9d13-e853c115d1c8","Type":"ContainerStarted","Data":"3e45891c4c7efe2911b9dc9b1ecc8b705d79afa569e5760003355d0bf297348b"} Oct 07 23:53:46 crc kubenswrapper[4871]: I1007 23:53:46.124090 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" event={"ID":"caf90bb9-05c5-40da-be2d-b1c57ef8075f","Type":"ContainerStarted","Data":"983218b295f7e5ead01adcdf72d6e2d64828c4618f82989392b8ea7512666163"} Oct 07 23:53:46 crc kubenswrapper[4871]: I1007 23:53:46.127533 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" event={"ID":"1d3e2b5e-61e3-45ec-b39b-ec6aee52c207","Type":"ContainerStarted","Data":"1f06044aac01bab337b77c56482d8912f11444119f97c0d056a80815f751d18e"} Oct 07 23:53:46 crc kubenswrapper[4871]: I1007 23:53:46.160162 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-hrq9c" podStartSLOduration=2.812304149 podStartE2EDuration="5.160140541s" podCreationTimestamp="2025-10-07 23:53:41 +0000 UTC" firstStartedPulling="2025-10-07 23:53:42.485853337 +0000 UTC m=+6296.288551410" lastFinishedPulling="2025-10-07 23:53:44.833689739 +0000 UTC m=+6298.636387802" observedRunningTime="2025-10-07 23:53:46.147271113 +0000 UTC m=+6299.949969206" watchObservedRunningTime="2025-10-07 23:53:46.160140541 +0000 UTC m=+6299.962838624" Oct 07 23:53:46 crc kubenswrapper[4871]: I1007 23:53:46.179659 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8497998f8-ggr6m" podStartSLOduration=2.838259189 podStartE2EDuration="5.179637852s" podCreationTimestamp="2025-10-07 23:53:41 +0000 UTC" firstStartedPulling="2025-10-07 23:53:42.485564389 +0000 UTC m=+6296.288262462" lastFinishedPulling="2025-10-07 23:53:44.826943052 +0000 UTC m=+6298.629641125" observedRunningTime="2025-10-07 23:53:46.16430334 +0000 UTC m=+6299.967001423" watchObservedRunningTime="2025-10-07 23:53:46.179637852 +0000 UTC m=+6299.982335925" Oct 07 23:53:50 crc kubenswrapper[4871]: I1007 23:53:50.218574 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" event={"ID":"07703b22-45ab-4158-9afe-a11f5f0f1e41","Type":"ContainerStarted","Data":"71fad6df83ec40d99a83f146019c2a731ddc7b3b1015aae3996da082a5237cd7"} Oct 07 23:53:50 crc kubenswrapper[4871]: I1007 23:53:50.219144 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:50 crc kubenswrapper[4871]: I1007 23:53:50.227724 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" Oct 07 23:53:50 crc kubenswrapper[4871]: I1007 23:53:50.234285 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" event={"ID":"48ff787b-84fa-4f90-9d13-e853c115d1c8","Type":"ContainerStarted","Data":"363b98ca70b07ac343190842339db6a077644ea3f7cc7ed26483d2193e54a4d0"} Oct 07 23:53:50 crc kubenswrapper[4871]: I1007 23:53:50.234400 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:53:50 crc kubenswrapper[4871]: I1007 23:53:50.240285 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-nr674" event={"ID":"a3adbb28-7bd8-418c-94e4-e0a1ff7c8726","Type":"ContainerStarted","Data":"4bc006bfbb1320bc9a2fdf1b8b1a59b5861e39ff24ea6faab3159e73ff31a618"} Oct 07 23:53:50 crc kubenswrapper[4871]: I1007 23:53:50.252982 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-ckv6s" podStartSLOduration=2.560845182 podStartE2EDuration="9.252962141s" podCreationTimestamp="2025-10-07 23:53:41 +0000 UTC" firstStartedPulling="2025-10-07 23:53:42.762074582 +0000 UTC m=+6296.564772655" lastFinishedPulling="2025-10-07 23:53:49.454191541 +0000 UTC m=+6303.256889614" observedRunningTime="2025-10-07 23:53:50.245705171 +0000 UTC m=+6304.048403244" watchObservedRunningTime="2025-10-07 23:53:50.252962141 +0000 UTC m=+6304.055660214" Oct 07 23:53:50 crc kubenswrapper[4871]: I1007 23:53:50.296566 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-nr674" podStartSLOduration=2.543858407 podStartE2EDuration="9.296551495s" podCreationTimestamp="2025-10-07 23:53:41 +0000 UTC" firstStartedPulling="2025-10-07 23:53:42.670262134 +0000 UTC m=+6296.472960207" lastFinishedPulling="2025-10-07 23:53:49.422955222 +0000 UTC m=+6303.225653295" observedRunningTime="2025-10-07 23:53:50.275232585 +0000 UTC m=+6304.077930658" watchObservedRunningTime="2025-10-07 23:53:50.296551495 +0000 UTC m=+6304.099249558" Oct 07 23:53:50 crc kubenswrapper[4871]: I1007 23:53:50.298819 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" podStartSLOduration=2.68541294 podStartE2EDuration="9.298812754s" podCreationTimestamp="2025-10-07 23:53:41 +0000 UTC" firstStartedPulling="2025-10-07 23:53:42.748797504 +0000 UTC m=+6296.551495577" lastFinishedPulling="2025-10-07 23:53:49.362197318 +0000 UTC m=+6303.164895391" observedRunningTime="2025-10-07 23:53:50.290522046 +0000 UTC m=+6304.093220119" watchObservedRunningTime="2025-10-07 23:53:50.298812754 +0000 UTC m=+6304.101510827" Oct 07 23:53:56 crc kubenswrapper[4871]: I1007 23:53:56.043118 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdkfc"] Oct 07 23:53:56 crc kubenswrapper[4871]: I1007 23:53:56.055204 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdkfc"] Oct 07 23:53:56 crc kubenswrapper[4871]: I1007 23:53:56.997062 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae15ef8e-e5a8-48ae-9317-5b0bbd197508" path="/var/lib/kubelet/pods/ae15ef8e-e5a8-48ae-9317-5b0bbd197508/volumes" Oct 07 23:53:57 crc kubenswrapper[4871]: I1007 23:53:57.023365 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zpwpr"] Oct 07 23:53:57 crc kubenswrapper[4871]: I1007 23:53:57.034952 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zpwpr"] Oct 07 23:53:58 crc kubenswrapper[4871]: I1007 23:53:58.994373 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81" path="/var/lib/kubelet/pods/e5e8b8e0-31d8-4c3a-a9e1-13f8f8d76d81/volumes" Oct 07 23:53:59 crc kubenswrapper[4871]: I1007 23:53:59.602827 4871 scope.go:117] "RemoveContainer" containerID="8d1bf31d634b04a296616b1251e9785c15f4a8d5b038e27a0dd45a18ba86a556" Oct 07 23:53:59 crc kubenswrapper[4871]: I1007 23:53:59.655527 4871 scope.go:117] "RemoveContainer" containerID="774c9bb65e51aeb69df62b157c02aa8bbd666899351e7f478ed28c8073822c62" Oct 07 23:53:59 crc kubenswrapper[4871]: I1007 23:53:59.697275 4871 scope.go:117] "RemoveContainer" containerID="3d127be49fc4e793064677b4ab32c550196edf416946fc832970b8a32dd8effd" Oct 07 23:53:59 crc kubenswrapper[4871]: I1007 23:53:59.751667 4871 scope.go:117] "RemoveContainer" containerID="9d2f7ea84742434bfbe6e84436cbd48d6a351ac96411bb9f8f7a5656404a1b7e" Oct 07 23:53:59 crc kubenswrapper[4871]: I1007 23:53:59.792310 4871 scope.go:117] "RemoveContainer" containerID="35e0ab42533abbf5426d11d2001a272cd9a71eb7377859f377db8c5d1addeca9" Oct 07 23:53:59 crc kubenswrapper[4871]: I1007 23:53:59.857841 4871 scope.go:117] "RemoveContainer" containerID="f48675f0e35395a85e4947dc21ed550901f70bd64fb64f4b4fd549f4b6ee63d1" Oct 07 23:53:59 crc kubenswrapper[4871]: I1007 23:53:59.877175 4871 scope.go:117] "RemoveContainer" containerID="0c1410f98ef3030ed6237b7719b3e4aaaed7b574b2c55523e5d535c73c9c67ab" Oct 07 23:53:59 crc kubenswrapper[4871]: I1007 23:53:59.929141 4871 scope.go:117] "RemoveContainer" containerID="3af5e9f9da3b45e938d3a8b0680b953996e731a5d83a7608ac7bd49fd9cfba11" Oct 07 23:53:59 crc kubenswrapper[4871]: I1007 23:53:59.950975 4871 scope.go:117] "RemoveContainer" containerID="47ae18d3fcf146c16f880e5af6617bf60cbfb09b064b3acc3aeb1ee071d589c7" Oct 07 23:54:02 crc kubenswrapper[4871]: I1007 23:54:02.180449 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-cjmdr" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.397509 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.398147 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="15bced8d-616c-4f9b-b265-b0eb216718c8" containerName="openstackclient" containerID="cri-o://3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e" gracePeriod=2 Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.411856 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.501334 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 07 23:54:04 crc kubenswrapper[4871]: E1007 23:54:04.501874 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15bced8d-616c-4f9b-b265-b0eb216718c8" containerName="openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.501899 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="15bced8d-616c-4f9b-b265-b0eb216718c8" containerName="openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.502108 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="15bced8d-616c-4f9b-b265-b0eb216718c8" containerName="openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.506437 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.540333 4871 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="15bced8d-616c-4f9b-b265-b0eb216718c8" podUID="64093265-fc84-469c-bbc0-55b87680c9e9" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.544993 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.637081 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx59m\" (UniqueName: \"kubernetes.io/projected/64093265-fc84-469c-bbc0-55b87680c9e9-kube-api-access-wx59m\") pod \"openstackclient\" (UID: \"64093265-fc84-469c-bbc0-55b87680c9e9\") " pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.637473 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/64093265-fc84-469c-bbc0-55b87680c9e9-openstack-config-secret\") pod \"openstackclient\" (UID: \"64093265-fc84-469c-bbc0-55b87680c9e9\") " pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.637618 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/64093265-fc84-469c-bbc0-55b87680c9e9-openstack-config\") pod \"openstackclient\" (UID: \"64093265-fc84-469c-bbc0-55b87680c9e9\") " pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.739463 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/64093265-fc84-469c-bbc0-55b87680c9e9-openstack-config\") pod \"openstackclient\" (UID: \"64093265-fc84-469c-bbc0-55b87680c9e9\") " pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.740103 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx59m\" (UniqueName: \"kubernetes.io/projected/64093265-fc84-469c-bbc0-55b87680c9e9-kube-api-access-wx59m\") pod \"openstackclient\" (UID: \"64093265-fc84-469c-bbc0-55b87680c9e9\") " pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.740204 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/64093265-fc84-469c-bbc0-55b87680c9e9-openstack-config-secret\") pod \"openstackclient\" (UID: \"64093265-fc84-469c-bbc0-55b87680c9e9\") " pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.740373 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/64093265-fc84-469c-bbc0-55b87680c9e9-openstack-config\") pod \"openstackclient\" (UID: \"64093265-fc84-469c-bbc0-55b87680c9e9\") " pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.743232 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.744546 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.746531 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/64093265-fc84-469c-bbc0-55b87680c9e9-openstack-config-secret\") pod \"openstackclient\" (UID: \"64093265-fc84-469c-bbc0-55b87680c9e9\") " pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.748283 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-vsxgz" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.754782 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.778017 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx59m\" (UniqueName: \"kubernetes.io/projected/64093265-fc84-469c-bbc0-55b87680c9e9-kube-api-access-wx59m\") pod \"openstackclient\" (UID: \"64093265-fc84-469c-bbc0-55b87680c9e9\") " pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.850340 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.853951 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq8bk\" (UniqueName: \"kubernetes.io/projected/96e5e7ae-f05f-4645-bd64-d51333e8bc4e-kube-api-access-lq8bk\") pod \"kube-state-metrics-0\" (UID: \"96e5e7ae-f05f-4645-bd64-d51333e8bc4e\") " pod="openstack/kube-state-metrics-0" Oct 07 23:54:04 crc kubenswrapper[4871]: I1007 23:54:04.958063 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq8bk\" (UniqueName: \"kubernetes.io/projected/96e5e7ae-f05f-4645-bd64-d51333e8bc4e-kube-api-access-lq8bk\") pod \"kube-state-metrics-0\" (UID: \"96e5e7ae-f05f-4645-bd64-d51333e8bc4e\") " pod="openstack/kube-state-metrics-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.021296 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq8bk\" (UniqueName: \"kubernetes.io/projected/96e5e7ae-f05f-4645-bd64-d51333e8bc4e-kube-api-access-lq8bk\") pod \"kube-state-metrics-0\" (UID: \"96e5e7ae-f05f-4645-bd64-d51333e8bc4e\") " pod="openstack/kube-state-metrics-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.266344 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.406676 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.413050 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.424346 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.424562 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.424668 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.426322 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-jc2qr" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.452893 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.482486 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ed4402de-a3a0-4c54-9675-a0766947a554-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.482644 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ed4402de-a3a0-4c54-9675-a0766947a554-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.482871 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bgr5\" (UniqueName: \"kubernetes.io/projected/ed4402de-a3a0-4c54-9675-a0766947a554-kube-api-access-6bgr5\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.483015 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ed4402de-a3a0-4c54-9675-a0766947a554-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.483077 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ed4402de-a3a0-4c54-9675-a0766947a554-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.483155 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ed4402de-a3a0-4c54-9675-a0766947a554-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.513935 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.513983 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.585014 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ed4402de-a3a0-4c54-9675-a0766947a554-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.585084 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ed4402de-a3a0-4c54-9675-a0766947a554-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.585122 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ed4402de-a3a0-4c54-9675-a0766947a554-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.585191 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bgr5\" (UniqueName: \"kubernetes.io/projected/ed4402de-a3a0-4c54-9675-a0766947a554-kube-api-access-6bgr5\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.585244 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ed4402de-a3a0-4c54-9675-a0766947a554-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.585277 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ed4402de-a3a0-4c54-9675-a0766947a554-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.585783 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ed4402de-a3a0-4c54-9675-a0766947a554-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.592576 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ed4402de-a3a0-4c54-9675-a0766947a554-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.592871 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ed4402de-a3a0-4c54-9675-a0766947a554-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.597290 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ed4402de-a3a0-4c54-9675-a0766947a554-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.597719 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ed4402de-a3a0-4c54-9675-a0766947a554-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.610430 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bgr5\" (UniqueName: \"kubernetes.io/projected/ed4402de-a3a0-4c54-9675-a0766947a554-kube-api-access-6bgr5\") pod \"alertmanager-metric-storage-0\" (UID: \"ed4402de-a3a0-4c54-9675-a0766947a554\") " pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.706557 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 23:54:05 crc kubenswrapper[4871]: I1007 23:54:05.755500 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.007804 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.058876 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.062647 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.068435 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-6fgdz" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.068599 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.068713 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.068835 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.070526 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.082105 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.104477 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.119735 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.200267 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-58697c15-5576-4031-99f3-b732b74396fa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58697c15-5576-4031-99f3-b732b74396fa\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.200299 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-config\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.200325 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.200355 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.200400 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.200434 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9mrq\" (UniqueName: \"kubernetes.io/projected/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-kube-api-access-b9mrq\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.200478 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.200523 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.302931 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.303236 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-config\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.303258 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-58697c15-5576-4031-99f3-b732b74396fa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58697c15-5576-4031-99f3-b732b74396fa\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.303285 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.303325 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.303364 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.303402 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9mrq\" (UniqueName: \"kubernetes.io/projected/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-kube-api-access-b9mrq\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.303445 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.304038 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.315689 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.317082 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-config\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.343264 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.344301 4871 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.344388 4871 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-58697c15-5576-4031-99f3-b732b74396fa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58697c15-5576-4031-99f3-b732b74396fa\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ddb8ff9c394ff5998d24a936efa097e828bd1923b8fc995d64b409af0a3c08c9/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.351608 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.353608 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.371184 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9mrq\" (UniqueName: \"kubernetes.io/projected/9edd958f-36cf-453d-a0f8-c5f27cdfdc07-kube-api-access-b9mrq\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.442766 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"96e5e7ae-f05f-4645-bd64-d51333e8bc4e","Type":"ContainerStarted","Data":"db5a81df7dd4a8766b6c789b9428454a0bd3a1597a396e44d1613066575d63ad"} Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.449068 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"64093265-fc84-469c-bbc0-55b87680c9e9","Type":"ContainerStarted","Data":"50159c52cbd508ee98add147e177c497ca833b88e6621908f4aa734f3e09108d"} Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.494827 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-58697c15-5576-4031-99f3-b732b74396fa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58697c15-5576-4031-99f3-b732b74396fa\") pod \"prometheus-metric-storage-0\" (UID: \"9edd958f-36cf-453d-a0f8-c5f27cdfdc07\") " pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.546314 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:06 crc kubenswrapper[4871]: I1007 23:54:06.693241 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 07 23:54:06 crc kubenswrapper[4871]: W1007 23:54:06.708509 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded4402de_a3a0_4c54_9675_a0766947a554.slice/crio-0c1c17104f7078eeae68cf7b8e9038f4e38d0bbf43293e8dd9c51f4badb7fb3f WatchSource:0}: Error finding container 0c1c17104f7078eeae68cf7b8e9038f4e38d0bbf43293e8dd9c51f4badb7fb3f: Status 404 returned error can't find the container with id 0c1c17104f7078eeae68cf7b8e9038f4e38d0bbf43293e8dd9c51f4badb7fb3f Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.115941 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.118671 4871 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="15bced8d-616c-4f9b-b265-b0eb216718c8" podUID="64093265-fc84-469c-bbc0-55b87680c9e9" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.143614 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.244123 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jglz\" (UniqueName: \"kubernetes.io/projected/15bced8d-616c-4f9b-b265-b0eb216718c8-kube-api-access-7jglz\") pod \"15bced8d-616c-4f9b-b265-b0eb216718c8\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.244483 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config-secret\") pod \"15bced8d-616c-4f9b-b265-b0eb216718c8\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.244526 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config\") pod \"15bced8d-616c-4f9b-b265-b0eb216718c8\" (UID: \"15bced8d-616c-4f9b-b265-b0eb216718c8\") " Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.250981 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15bced8d-616c-4f9b-b265-b0eb216718c8-kube-api-access-7jglz" (OuterVolumeSpecName: "kube-api-access-7jglz") pod "15bced8d-616c-4f9b-b265-b0eb216718c8" (UID: "15bced8d-616c-4f9b-b265-b0eb216718c8"). InnerVolumeSpecName "kube-api-access-7jglz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.306853 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "15bced8d-616c-4f9b-b265-b0eb216718c8" (UID: "15bced8d-616c-4f9b-b265-b0eb216718c8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.313585 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "15bced8d-616c-4f9b-b265-b0eb216718c8" (UID: "15bced8d-616c-4f9b-b265-b0eb216718c8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.347773 4871 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.347829 4871 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/15bced8d-616c-4f9b-b265-b0eb216718c8-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.347842 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jglz\" (UniqueName: \"kubernetes.io/projected/15bced8d-616c-4f9b-b265-b0eb216718c8-kube-api-access-7jglz\") on node \"crc\" DevicePath \"\"" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.459575 4871 generic.go:334] "Generic (PLEG): container finished" podID="15bced8d-616c-4f9b-b265-b0eb216718c8" containerID="3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e" exitCode=137 Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.459660 4871 scope.go:117] "RemoveContainer" containerID="3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.461003 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.461134 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"96e5e7ae-f05f-4645-bd64-d51333e8bc4e","Type":"ContainerStarted","Data":"e97a24370a783f0dd8be5d52a6fc38e93ffbf31178d335ba5f36bb9a8089f177"} Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.463836 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.465167 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"64093265-fc84-469c-bbc0-55b87680c9e9","Type":"ContainerStarted","Data":"335816b417d1054f1a47f03cbfc3b9bfb078d7c1335bb672a91de08fb6d1ad67"} Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.469420 4871 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="15bced8d-616c-4f9b-b265-b0eb216718c8" podUID="64093265-fc84-469c-bbc0-55b87680c9e9" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.473445 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ed4402de-a3a0-4c54-9675-a0766947a554","Type":"ContainerStarted","Data":"0c1c17104f7078eeae68cf7b8e9038f4e38d0bbf43293e8dd9c51f4badb7fb3f"} Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.477425 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9edd958f-36cf-453d-a0f8-c5f27cdfdc07","Type":"ContainerStarted","Data":"54653209f41b35a45baafcceb99694af555f5411b8bfacff84934986f49b8a7b"} Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.491658 4871 scope.go:117] "RemoveContainer" containerID="3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e" Oct 07 23:54:07 crc kubenswrapper[4871]: E1007 23:54:07.492264 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e\": container with ID starting with 3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e not found: ID does not exist" containerID="3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.492331 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e"} err="failed to get container status \"3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e\": rpc error: code = NotFound desc = could not find container \"3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e\": container with ID starting with 3b3490680bfb195b9cd8005b38776255a1bc8ed4bc4b946823166a1a9defe31e not found: ID does not exist" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.495176 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.954802448 podStartE2EDuration="3.495158251s" podCreationTimestamp="2025-10-07 23:54:04 +0000 UTC" firstStartedPulling="2025-10-07 23:54:06.119539529 +0000 UTC m=+6319.922237602" lastFinishedPulling="2025-10-07 23:54:06.659895332 +0000 UTC m=+6320.462593405" observedRunningTime="2025-10-07 23:54:07.490786336 +0000 UTC m=+6321.293484419" watchObservedRunningTime="2025-10-07 23:54:07.495158251 +0000 UTC m=+6321.297856334" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.515978 4871 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="15bced8d-616c-4f9b-b265-b0eb216718c8" podUID="64093265-fc84-469c-bbc0-55b87680c9e9" Oct 07 23:54:07 crc kubenswrapper[4871]: I1007 23:54:07.523425 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.523405452 podStartE2EDuration="3.523405452s" podCreationTimestamp="2025-10-07 23:54:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:54:07.513353638 +0000 UTC m=+6321.316051711" watchObservedRunningTime="2025-10-07 23:54:07.523405452 +0000 UTC m=+6321.326103525" Oct 07 23:54:09 crc kubenswrapper[4871]: I1007 23:54:08.998372 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15bced8d-616c-4f9b-b265-b0eb216718c8" path="/var/lib/kubelet/pods/15bced8d-616c-4f9b-b265-b0eb216718c8/volumes" Oct 07 23:54:10 crc kubenswrapper[4871]: I1007 23:54:10.043357 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-m5lll"] Oct 07 23:54:10 crc kubenswrapper[4871]: I1007 23:54:10.060997 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-m5lll"] Oct 07 23:54:10 crc kubenswrapper[4871]: I1007 23:54:10.994456 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcae407f-430e-49ad-833e-c027eeabb01a" path="/var/lib/kubelet/pods/bcae407f-430e-49ad-833e-c027eeabb01a/volumes" Oct 07 23:54:13 crc kubenswrapper[4871]: I1007 23:54:13.541783 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ed4402de-a3a0-4c54-9675-a0766947a554","Type":"ContainerStarted","Data":"1243ae33df96df57512ed64719b77dcd63f6dcccdabc82b20935f50c6764d710"} Oct 07 23:54:13 crc kubenswrapper[4871]: I1007 23:54:13.544983 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9edd958f-36cf-453d-a0f8-c5f27cdfdc07","Type":"ContainerStarted","Data":"96eab48b9ab51064ea86080e90688b98c06da2656e3daa296a862f0fb7e46a8a"} Oct 07 23:54:15 crc kubenswrapper[4871]: I1007 23:54:15.273700 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 07 23:54:21 crc kubenswrapper[4871]: I1007 23:54:21.633658 4871 generic.go:334] "Generic (PLEG): container finished" podID="ed4402de-a3a0-4c54-9675-a0766947a554" containerID="1243ae33df96df57512ed64719b77dcd63f6dcccdabc82b20935f50c6764d710" exitCode=0 Oct 07 23:54:21 crc kubenswrapper[4871]: I1007 23:54:21.633729 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ed4402de-a3a0-4c54-9675-a0766947a554","Type":"ContainerDied","Data":"1243ae33df96df57512ed64719b77dcd63f6dcccdabc82b20935f50c6764d710"} Oct 07 23:54:22 crc kubenswrapper[4871]: I1007 23:54:22.662553 4871 generic.go:334] "Generic (PLEG): container finished" podID="9edd958f-36cf-453d-a0f8-c5f27cdfdc07" containerID="96eab48b9ab51064ea86080e90688b98c06da2656e3daa296a862f0fb7e46a8a" exitCode=0 Oct 07 23:54:22 crc kubenswrapper[4871]: I1007 23:54:22.662615 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9edd958f-36cf-453d-a0f8-c5f27cdfdc07","Type":"ContainerDied","Data":"96eab48b9ab51064ea86080e90688b98c06da2656e3daa296a862f0fb7e46a8a"} Oct 07 23:54:24 crc kubenswrapper[4871]: I1007 23:54:24.689504 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ed4402de-a3a0-4c54-9675-a0766947a554","Type":"ContainerStarted","Data":"f4da87dfff4faa8cb9ca412b15a4e72b8eeab15a5e2f049326ec28f4c0229ce5"} Oct 07 23:54:28 crc kubenswrapper[4871]: I1007 23:54:28.738464 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ed4402de-a3a0-4c54-9675-a0766947a554","Type":"ContainerStarted","Data":"ea911c128e9d20e4bf38dea61f5eb858e90a2ce5df0e633e200da062eacdbb6d"} Oct 07 23:54:28 crc kubenswrapper[4871]: I1007 23:54:28.739288 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:28 crc kubenswrapper[4871]: I1007 23:54:28.746646 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 07 23:54:28 crc kubenswrapper[4871]: I1007 23:54:28.787018 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.721817873 podStartE2EDuration="23.786996879s" podCreationTimestamp="2025-10-07 23:54:05 +0000 UTC" firstStartedPulling="2025-10-07 23:54:06.711885996 +0000 UTC m=+6320.514584069" lastFinishedPulling="2025-10-07 23:54:23.777065002 +0000 UTC m=+6337.579763075" observedRunningTime="2025-10-07 23:54:28.776906394 +0000 UTC m=+6342.579604547" watchObservedRunningTime="2025-10-07 23:54:28.786996879 +0000 UTC m=+6342.589694972" Oct 07 23:54:29 crc kubenswrapper[4871]: I1007 23:54:29.762834 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9edd958f-36cf-453d-a0f8-c5f27cdfdc07","Type":"ContainerStarted","Data":"5d63c43ab89305ca29d348735b91cb8c1db5e9d38ca0f15d7ea2792b619dd9ab"} Oct 07 23:54:33 crc kubenswrapper[4871]: I1007 23:54:33.833987 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9edd958f-36cf-453d-a0f8-c5f27cdfdc07","Type":"ContainerStarted","Data":"48514458e9b8a8a1a2124d4dcb075278fdeda6c307e60ff1497e034efa7d2420"} Oct 07 23:54:35 crc kubenswrapper[4871]: I1007 23:54:35.512263 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:54:35 crc kubenswrapper[4871]: I1007 23:54:35.512553 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:54:36 crc kubenswrapper[4871]: I1007 23:54:36.872883 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9edd958f-36cf-453d-a0f8-c5f27cdfdc07","Type":"ContainerStarted","Data":"3ce93192ab3be46e9c51b81281ec21960619254b8582723107daba4fcec59b5c"} Oct 07 23:54:36 crc kubenswrapper[4871]: I1007 23:54:36.916914 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.829785187 podStartE2EDuration="32.916876928s" podCreationTimestamp="2025-10-07 23:54:04 +0000 UTC" firstStartedPulling="2025-10-07 23:54:07.152581975 +0000 UTC m=+6320.955280048" lastFinishedPulling="2025-10-07 23:54:36.239673716 +0000 UTC m=+6350.042371789" observedRunningTime="2025-10-07 23:54:36.90550872 +0000 UTC m=+6350.708206843" watchObservedRunningTime="2025-10-07 23:54:36.916876928 +0000 UTC m=+6350.719575041" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.307061 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.312232 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.315695 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.316322 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.334928 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.463508 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbdr8\" (UniqueName: \"kubernetes.io/projected/0a81fdaf-895c-415c-b921-6bd563b3eeb5-kube-api-access-gbdr8\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.464009 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-run-httpd\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.464171 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-log-httpd\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.464413 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.464722 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.464946 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-scripts\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.465151 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-config-data\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.547528 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.567376 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbdr8\" (UniqueName: \"kubernetes.io/projected/0a81fdaf-895c-415c-b921-6bd563b3eeb5-kube-api-access-gbdr8\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.567458 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-run-httpd\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.567486 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-log-httpd\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.567506 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.567577 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.567607 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-scripts\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.567632 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-config-data\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.568387 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-log-httpd\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.568380 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-run-httpd\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.576707 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.579914 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-scripts\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.582758 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.583517 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-config-data\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.587464 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbdr8\" (UniqueName: \"kubernetes.io/projected/0a81fdaf-895c-415c-b921-6bd563b3eeb5-kube-api-access-gbdr8\") pod \"ceilometer-0\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " pod="openstack/ceilometer-0" Oct 07 23:54:41 crc kubenswrapper[4871]: I1007 23:54:41.649166 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:54:42 crc kubenswrapper[4871]: I1007 23:54:42.269764 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:54:42 crc kubenswrapper[4871]: I1007 23:54:42.997147 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerStarted","Data":"ab9b0dd0b7bf8eb4288cd15b518235ae0ce02048ac711596e3bbf55080ed09e4"} Oct 07 23:54:44 crc kubenswrapper[4871]: I1007 23:54:44.006655 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerStarted","Data":"7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b"} Oct 07 23:54:44 crc kubenswrapper[4871]: I1007 23:54:44.007205 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerStarted","Data":"90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea"} Oct 07 23:54:45 crc kubenswrapper[4871]: I1007 23:54:45.020125 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerStarted","Data":"2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21"} Oct 07 23:54:45 crc kubenswrapper[4871]: I1007 23:54:45.902486 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9j9pb"] Oct 07 23:54:45 crc kubenswrapper[4871]: I1007 23:54:45.905456 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:45 crc kubenswrapper[4871]: I1007 23:54:45.914306 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9j9pb"] Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.069174 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-catalog-content\") pod \"certified-operators-9j9pb\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.071710 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-utilities\") pod \"certified-operators-9j9pb\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.071817 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv7cr\" (UniqueName: \"kubernetes.io/projected/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-kube-api-access-zv7cr\") pod \"certified-operators-9j9pb\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.178234 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-catalog-content\") pod \"certified-operators-9j9pb\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.178292 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-utilities\") pod \"certified-operators-9j9pb\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.178322 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv7cr\" (UniqueName: \"kubernetes.io/projected/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-kube-api-access-zv7cr\") pod \"certified-operators-9j9pb\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.179361 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-catalog-content\") pod \"certified-operators-9j9pb\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.179583 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-utilities\") pod \"certified-operators-9j9pb\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.220622 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv7cr\" (UniqueName: \"kubernetes.io/projected/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-kube-api-access-zv7cr\") pod \"certified-operators-9j9pb\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.414240 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:46 crc kubenswrapper[4871]: W1007 23:54:46.977456 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9f2320a_4618_468c_8b7d_2c8bb93cbe79.slice/crio-4b46cccac396ca1c81713599854c3a7ddd8ef0f5168c549b232c5b5fd6553053 WatchSource:0}: Error finding container 4b46cccac396ca1c81713599854c3a7ddd8ef0f5168c549b232c5b5fd6553053: Status 404 returned error can't find the container with id 4b46cccac396ca1c81713599854c3a7ddd8ef0f5168c549b232c5b5fd6553053 Oct 07 23:54:46 crc kubenswrapper[4871]: I1007 23:54:46.978466 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9j9pb"] Oct 07 23:54:47 crc kubenswrapper[4871]: I1007 23:54:47.040657 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerStarted","Data":"99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193"} Oct 07 23:54:47 crc kubenswrapper[4871]: I1007 23:54:47.040845 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 23:54:47 crc kubenswrapper[4871]: I1007 23:54:47.041712 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9j9pb" event={"ID":"a9f2320a-4618-468c-8b7d-2c8bb93cbe79","Type":"ContainerStarted","Data":"4b46cccac396ca1c81713599854c3a7ddd8ef0f5168c549b232c5b5fd6553053"} Oct 07 23:54:48 crc kubenswrapper[4871]: I1007 23:54:48.053484 4871 generic.go:334] "Generic (PLEG): container finished" podID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerID="ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f" exitCode=0 Oct 07 23:54:48 crc kubenswrapper[4871]: I1007 23:54:48.053558 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9j9pb" event={"ID":"a9f2320a-4618-468c-8b7d-2c8bb93cbe79","Type":"ContainerDied","Data":"ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f"} Oct 07 23:54:48 crc kubenswrapper[4871]: I1007 23:54:48.077131 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.471265054 podStartE2EDuration="7.077112922s" podCreationTimestamp="2025-10-07 23:54:41 +0000 UTC" firstStartedPulling="2025-10-07 23:54:42.298206257 +0000 UTC m=+6356.100904330" lastFinishedPulling="2025-10-07 23:54:45.904054105 +0000 UTC m=+6359.706752198" observedRunningTime="2025-10-07 23:54:47.065395486 +0000 UTC m=+6360.868093559" watchObservedRunningTime="2025-10-07 23:54:48.077112922 +0000 UTC m=+6361.879811005" Oct 07 23:54:50 crc kubenswrapper[4871]: I1007 23:54:50.084435 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9j9pb" event={"ID":"a9f2320a-4618-468c-8b7d-2c8bb93cbe79","Type":"ContainerStarted","Data":"61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835"} Oct 07 23:54:50 crc kubenswrapper[4871]: I1007 23:54:50.285419 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-gbqcd"] Oct 07 23:54:50 crc kubenswrapper[4871]: I1007 23:54:50.286775 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-gbqcd" Oct 07 23:54:50 crc kubenswrapper[4871]: I1007 23:54:50.313028 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-gbqcd"] Oct 07 23:54:50 crc kubenswrapper[4871]: I1007 23:54:50.381510 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkb25\" (UniqueName: \"kubernetes.io/projected/bb8d68cc-caee-453d-8e02-9104fdb415c9-kube-api-access-wkb25\") pod \"aodh-db-create-gbqcd\" (UID: \"bb8d68cc-caee-453d-8e02-9104fdb415c9\") " pod="openstack/aodh-db-create-gbqcd" Oct 07 23:54:50 crc kubenswrapper[4871]: I1007 23:54:50.489274 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkb25\" (UniqueName: \"kubernetes.io/projected/bb8d68cc-caee-453d-8e02-9104fdb415c9-kube-api-access-wkb25\") pod \"aodh-db-create-gbqcd\" (UID: \"bb8d68cc-caee-453d-8e02-9104fdb415c9\") " pod="openstack/aodh-db-create-gbqcd" Oct 07 23:54:50 crc kubenswrapper[4871]: I1007 23:54:50.528175 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkb25\" (UniqueName: \"kubernetes.io/projected/bb8d68cc-caee-453d-8e02-9104fdb415c9-kube-api-access-wkb25\") pod \"aodh-db-create-gbqcd\" (UID: \"bb8d68cc-caee-453d-8e02-9104fdb415c9\") " pod="openstack/aodh-db-create-gbqcd" Oct 07 23:54:50 crc kubenswrapper[4871]: I1007 23:54:50.645231 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-gbqcd" Oct 07 23:54:51 crc kubenswrapper[4871]: I1007 23:54:51.095489 4871 generic.go:334] "Generic (PLEG): container finished" podID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerID="61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835" exitCode=0 Oct 07 23:54:51 crc kubenswrapper[4871]: I1007 23:54:51.095550 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9j9pb" event={"ID":"a9f2320a-4618-468c-8b7d-2c8bb93cbe79","Type":"ContainerDied","Data":"61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835"} Oct 07 23:54:51 crc kubenswrapper[4871]: W1007 23:54:51.156223 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb8d68cc_caee_453d_8e02_9104fdb415c9.slice/crio-c7a66a729dc05a6b04484ba7d47aee9174f6ae6b8d3916165f980db2ec5c3590 WatchSource:0}: Error finding container c7a66a729dc05a6b04484ba7d47aee9174f6ae6b8d3916165f980db2ec5c3590: Status 404 returned error can't find the container with id c7a66a729dc05a6b04484ba7d47aee9174f6ae6b8d3916165f980db2ec5c3590 Oct 07 23:54:51 crc kubenswrapper[4871]: I1007 23:54:51.156223 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-gbqcd"] Oct 07 23:54:51 crc kubenswrapper[4871]: I1007 23:54:51.548198 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:51 crc kubenswrapper[4871]: I1007 23:54:51.553035 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:52 crc kubenswrapper[4871]: I1007 23:54:52.107009 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9j9pb" event={"ID":"a9f2320a-4618-468c-8b7d-2c8bb93cbe79","Type":"ContainerStarted","Data":"22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986"} Oct 07 23:54:52 crc kubenswrapper[4871]: I1007 23:54:52.109323 4871 generic.go:334] "Generic (PLEG): container finished" podID="bb8d68cc-caee-453d-8e02-9104fdb415c9" containerID="1b4388d08b3a7187b520e69babee3c406b7c083955e0797d655ccecbfc3666c6" exitCode=0 Oct 07 23:54:52 crc kubenswrapper[4871]: I1007 23:54:52.110926 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-gbqcd" event={"ID":"bb8d68cc-caee-453d-8e02-9104fdb415c9","Type":"ContainerDied","Data":"1b4388d08b3a7187b520e69babee3c406b7c083955e0797d655ccecbfc3666c6"} Oct 07 23:54:52 crc kubenswrapper[4871]: I1007 23:54:52.110982 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-gbqcd" event={"ID":"bb8d68cc-caee-453d-8e02-9104fdb415c9","Type":"ContainerStarted","Data":"c7a66a729dc05a6b04484ba7d47aee9174f6ae6b8d3916165f980db2ec5c3590"} Oct 07 23:54:52 crc kubenswrapper[4871]: I1007 23:54:52.112256 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 07 23:54:52 crc kubenswrapper[4871]: I1007 23:54:52.137055 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9j9pb" podStartSLOduration=3.675344404 podStartE2EDuration="7.137025431s" podCreationTimestamp="2025-10-07 23:54:45 +0000 UTC" firstStartedPulling="2025-10-07 23:54:48.056158693 +0000 UTC m=+6361.858856766" lastFinishedPulling="2025-10-07 23:54:51.51783972 +0000 UTC m=+6365.320537793" observedRunningTime="2025-10-07 23:54:52.123815514 +0000 UTC m=+6365.926513607" watchObservedRunningTime="2025-10-07 23:54:52.137025431 +0000 UTC m=+6365.939723524" Oct 07 23:54:53 crc kubenswrapper[4871]: I1007 23:54:53.538668 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-gbqcd" Oct 07 23:54:53 crc kubenswrapper[4871]: I1007 23:54:53.587929 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkb25\" (UniqueName: \"kubernetes.io/projected/bb8d68cc-caee-453d-8e02-9104fdb415c9-kube-api-access-wkb25\") pod \"bb8d68cc-caee-453d-8e02-9104fdb415c9\" (UID: \"bb8d68cc-caee-453d-8e02-9104fdb415c9\") " Oct 07 23:54:53 crc kubenswrapper[4871]: I1007 23:54:53.593540 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb8d68cc-caee-453d-8e02-9104fdb415c9-kube-api-access-wkb25" (OuterVolumeSpecName: "kube-api-access-wkb25") pod "bb8d68cc-caee-453d-8e02-9104fdb415c9" (UID: "bb8d68cc-caee-453d-8e02-9104fdb415c9"). InnerVolumeSpecName "kube-api-access-wkb25". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:54:53 crc kubenswrapper[4871]: I1007 23:54:53.690261 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkb25\" (UniqueName: \"kubernetes.io/projected/bb8d68cc-caee-453d-8e02-9104fdb415c9-kube-api-access-wkb25\") on node \"crc\" DevicePath \"\"" Oct 07 23:54:54 crc kubenswrapper[4871]: I1007 23:54:54.128772 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-gbqcd" event={"ID":"bb8d68cc-caee-453d-8e02-9104fdb415c9","Type":"ContainerDied","Data":"c7a66a729dc05a6b04484ba7d47aee9174f6ae6b8d3916165f980db2ec5c3590"} Oct 07 23:54:54 crc kubenswrapper[4871]: I1007 23:54:54.128834 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7a66a729dc05a6b04484ba7d47aee9174f6ae6b8d3916165f980db2ec5c3590" Oct 07 23:54:54 crc kubenswrapper[4871]: I1007 23:54:54.128887 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-gbqcd" Oct 07 23:54:56 crc kubenswrapper[4871]: I1007 23:54:56.026630 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-gtk4b"] Oct 07 23:54:56 crc kubenswrapper[4871]: I1007 23:54:56.036383 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-gtk4b"] Oct 07 23:54:56 crc kubenswrapper[4871]: I1007 23:54:56.414474 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:56 crc kubenswrapper[4871]: I1007 23:54:56.414595 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:56 crc kubenswrapper[4871]: I1007 23:54:56.501718 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:57 crc kubenswrapper[4871]: I1007 23:54:57.002633 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d021d8ab-b7ae-4136-a946-2729bdd971b9" path="/var/lib/kubelet/pods/d021d8ab-b7ae-4136-a946-2729bdd971b9/volumes" Oct 07 23:54:57 crc kubenswrapper[4871]: I1007 23:54:57.266332 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:57 crc kubenswrapper[4871]: I1007 23:54:57.317669 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9j9pb"] Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.211220 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9j9pb" podUID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerName="registry-server" containerID="cri-o://22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986" gracePeriod=2 Oct 07 23:54:59 crc kubenswrapper[4871]: E1007 23:54:59.498124 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9f2320a_4618_468c_8b7d_2c8bb93cbe79.slice/crio-22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986.scope\": RecentStats: unable to find data in memory cache]" Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.734423 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.828832 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv7cr\" (UniqueName: \"kubernetes.io/projected/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-kube-api-access-zv7cr\") pod \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.828941 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-utilities\") pod \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.829245 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-catalog-content\") pod \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\" (UID: \"a9f2320a-4618-468c-8b7d-2c8bb93cbe79\") " Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.829883 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-utilities" (OuterVolumeSpecName: "utilities") pod "a9f2320a-4618-468c-8b7d-2c8bb93cbe79" (UID: "a9f2320a-4618-468c-8b7d-2c8bb93cbe79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.834866 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-kube-api-access-zv7cr" (OuterVolumeSpecName: "kube-api-access-zv7cr") pod "a9f2320a-4618-468c-8b7d-2c8bb93cbe79" (UID: "a9f2320a-4618-468c-8b7d-2c8bb93cbe79"). InnerVolumeSpecName "kube-api-access-zv7cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.905526 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9f2320a-4618-468c-8b7d-2c8bb93cbe79" (UID: "a9f2320a-4618-468c-8b7d-2c8bb93cbe79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.939288 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.939479 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv7cr\" (UniqueName: \"kubernetes.io/projected/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-kube-api-access-zv7cr\") on node \"crc\" DevicePath \"\"" Oct 07 23:54:59 crc kubenswrapper[4871]: I1007 23:54:59.939539 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f2320a-4618-468c-8b7d-2c8bb93cbe79-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.182414 4871 scope.go:117] "RemoveContainer" containerID="4a2c2efd8c6101f6ff7df800e3a4a0a8b13b16ff72ae1ec43233a855fb874ea2" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.216782 4871 scope.go:117] "RemoveContainer" containerID="3104f2ccfd5f5891af2334afadb5fedde54e4afa72e94fea9cbff6941bc50f09" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.231590 4871 generic.go:334] "Generic (PLEG): container finished" podID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerID="22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986" exitCode=0 Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.231622 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9j9pb" event={"ID":"a9f2320a-4618-468c-8b7d-2c8bb93cbe79","Type":"ContainerDied","Data":"22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986"} Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.231655 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9j9pb" event={"ID":"a9f2320a-4618-468c-8b7d-2c8bb93cbe79","Type":"ContainerDied","Data":"4b46cccac396ca1c81713599854c3a7ddd8ef0f5168c549b232c5b5fd6553053"} Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.231683 4871 scope.go:117] "RemoveContainer" containerID="22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.231691 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9j9pb" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.303448 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9j9pb"] Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.311633 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9j9pb"] Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.315993 4871 scope.go:117] "RemoveContainer" containerID="61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.332195 4871 scope.go:117] "RemoveContainer" containerID="57214d8df072bd58d6bd487bbb798e194944cf9f95d2de5da11d404777dbe252" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.418325 4871 scope.go:117] "RemoveContainer" containerID="ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.485425 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-e5a4-account-create-dshgj"] Oct 07 23:55:00 crc kubenswrapper[4871]: E1007 23:55:00.486117 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerName="registry-server" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.486190 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerName="registry-server" Oct 07 23:55:00 crc kubenswrapper[4871]: E1007 23:55:00.486256 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb8d68cc-caee-453d-8e02-9104fdb415c9" containerName="mariadb-database-create" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.486305 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb8d68cc-caee-453d-8e02-9104fdb415c9" containerName="mariadb-database-create" Oct 07 23:55:00 crc kubenswrapper[4871]: E1007 23:55:00.486359 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerName="extract-content" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.486407 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerName="extract-content" Oct 07 23:55:00 crc kubenswrapper[4871]: E1007 23:55:00.486599 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerName="extract-utilities" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.486668 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerName="extract-utilities" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.486937 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb8d68cc-caee-453d-8e02-9104fdb415c9" containerName="mariadb-database-create" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.488914 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" containerName="registry-server" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.489730 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-e5a4-account-create-dshgj" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.492187 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.503762 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-e5a4-account-create-dshgj"] Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.515172 4871 scope.go:117] "RemoveContainer" containerID="22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986" Oct 07 23:55:00 crc kubenswrapper[4871]: E1007 23:55:00.517492 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986\": container with ID starting with 22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986 not found: ID does not exist" containerID="22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.517523 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986"} err="failed to get container status \"22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986\": rpc error: code = NotFound desc = could not find container \"22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986\": container with ID starting with 22eefb0c5ceeeca9c4209faa6c02006544dbb7497428cf3e58818ef0fa924986 not found: ID does not exist" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.517562 4871 scope.go:117] "RemoveContainer" containerID="61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835" Oct 07 23:55:00 crc kubenswrapper[4871]: E1007 23:55:00.517869 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835\": container with ID starting with 61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835 not found: ID does not exist" containerID="61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.517973 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835"} err="failed to get container status \"61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835\": rpc error: code = NotFound desc = could not find container \"61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835\": container with ID starting with 61afedcdba3d18a99d597ea67354d8649d7403d299cb11395a67c5d80712a835 not found: ID does not exist" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.518039 4871 scope.go:117] "RemoveContainer" containerID="ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f" Oct 07 23:55:00 crc kubenswrapper[4871]: E1007 23:55:00.518426 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f\": container with ID starting with ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f not found: ID does not exist" containerID="ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.518445 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f"} err="failed to get container status \"ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f\": rpc error: code = NotFound desc = could not find container \"ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f\": container with ID starting with ee07ef5e5f7565903dcf46a9eb89f3f5996981d9e180d3803482667dcd11425f not found: ID does not exist" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.551661 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2lwg\" (UniqueName: \"kubernetes.io/projected/079ddc46-d11f-4903-9bad-5d0753a722cd-kube-api-access-q2lwg\") pod \"aodh-e5a4-account-create-dshgj\" (UID: \"079ddc46-d11f-4903-9bad-5d0753a722cd\") " pod="openstack/aodh-e5a4-account-create-dshgj" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.654599 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2lwg\" (UniqueName: \"kubernetes.io/projected/079ddc46-d11f-4903-9bad-5d0753a722cd-kube-api-access-q2lwg\") pod \"aodh-e5a4-account-create-dshgj\" (UID: \"079ddc46-d11f-4903-9bad-5d0753a722cd\") " pod="openstack/aodh-e5a4-account-create-dshgj" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.675376 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2lwg\" (UniqueName: \"kubernetes.io/projected/079ddc46-d11f-4903-9bad-5d0753a722cd-kube-api-access-q2lwg\") pod \"aodh-e5a4-account-create-dshgj\" (UID: \"079ddc46-d11f-4903-9bad-5d0753a722cd\") " pod="openstack/aodh-e5a4-account-create-dshgj" Oct 07 23:55:00 crc kubenswrapper[4871]: I1007 23:55:00.863772 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-e5a4-account-create-dshgj" Oct 07 23:55:01 crc kubenswrapper[4871]: I1007 23:55:01.000159 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9f2320a-4618-468c-8b7d-2c8bb93cbe79" path="/var/lib/kubelet/pods/a9f2320a-4618-468c-8b7d-2c8bb93cbe79/volumes" Oct 07 23:55:01 crc kubenswrapper[4871]: I1007 23:55:01.213849 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-e5a4-account-create-dshgj"] Oct 07 23:55:01 crc kubenswrapper[4871]: W1007 23:55:01.217020 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod079ddc46_d11f_4903_9bad_5d0753a722cd.slice/crio-f2d8a618e178acd76874ca380a9055d88d9b251fcd7d1ca5eb7495f40e9b63dd WatchSource:0}: Error finding container f2d8a618e178acd76874ca380a9055d88d9b251fcd7d1ca5eb7495f40e9b63dd: Status 404 returned error can't find the container with id f2d8a618e178acd76874ca380a9055d88d9b251fcd7d1ca5eb7495f40e9b63dd Oct 07 23:55:01 crc kubenswrapper[4871]: I1007 23:55:01.256508 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-e5a4-account-create-dshgj" event={"ID":"079ddc46-d11f-4903-9bad-5d0753a722cd","Type":"ContainerStarted","Data":"f2d8a618e178acd76874ca380a9055d88d9b251fcd7d1ca5eb7495f40e9b63dd"} Oct 07 23:55:02 crc kubenswrapper[4871]: I1007 23:55:02.268514 4871 generic.go:334] "Generic (PLEG): container finished" podID="079ddc46-d11f-4903-9bad-5d0753a722cd" containerID="79930ce111ddd8d4a05be42d44fc592b673016acdf7776e3ce1f6fd081079652" exitCode=0 Oct 07 23:55:02 crc kubenswrapper[4871]: I1007 23:55:02.268643 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-e5a4-account-create-dshgj" event={"ID":"079ddc46-d11f-4903-9bad-5d0753a722cd","Type":"ContainerDied","Data":"79930ce111ddd8d4a05be42d44fc592b673016acdf7776e3ce1f6fd081079652"} Oct 07 23:55:03 crc kubenswrapper[4871]: I1007 23:55:03.741588 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-e5a4-account-create-dshgj" Oct 07 23:55:03 crc kubenswrapper[4871]: I1007 23:55:03.839339 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2lwg\" (UniqueName: \"kubernetes.io/projected/079ddc46-d11f-4903-9bad-5d0753a722cd-kube-api-access-q2lwg\") pod \"079ddc46-d11f-4903-9bad-5d0753a722cd\" (UID: \"079ddc46-d11f-4903-9bad-5d0753a722cd\") " Oct 07 23:55:03 crc kubenswrapper[4871]: I1007 23:55:03.845468 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/079ddc46-d11f-4903-9bad-5d0753a722cd-kube-api-access-q2lwg" (OuterVolumeSpecName: "kube-api-access-q2lwg") pod "079ddc46-d11f-4903-9bad-5d0753a722cd" (UID: "079ddc46-d11f-4903-9bad-5d0753a722cd"). InnerVolumeSpecName "kube-api-access-q2lwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:55:03 crc kubenswrapper[4871]: I1007 23:55:03.941893 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2lwg\" (UniqueName: \"kubernetes.io/projected/079ddc46-d11f-4903-9bad-5d0753a722cd-kube-api-access-q2lwg\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:04 crc kubenswrapper[4871]: I1007 23:55:04.294632 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-e5a4-account-create-dshgj" event={"ID":"079ddc46-d11f-4903-9bad-5d0753a722cd","Type":"ContainerDied","Data":"f2d8a618e178acd76874ca380a9055d88d9b251fcd7d1ca5eb7495f40e9b63dd"} Oct 07 23:55:04 crc kubenswrapper[4871]: I1007 23:55:04.294932 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2d8a618e178acd76874ca380a9055d88d9b251fcd7d1ca5eb7495f40e9b63dd" Oct 07 23:55:04 crc kubenswrapper[4871]: I1007 23:55:04.294721 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-e5a4-account-create-dshgj" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.512543 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.512937 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.513001 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.514173 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.514271 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" gracePeriod=600 Oct 07 23:55:05 crc kubenswrapper[4871]: E1007 23:55:05.644624 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.919372 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-2m68z"] Oct 07 23:55:05 crc kubenswrapper[4871]: E1007 23:55:05.919866 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="079ddc46-d11f-4903-9bad-5d0753a722cd" containerName="mariadb-account-create" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.919879 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="079ddc46-d11f-4903-9bad-5d0753a722cd" containerName="mariadb-account-create" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.920098 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="079ddc46-d11f-4903-9bad-5d0753a722cd" containerName="mariadb-account-create" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.920904 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.922733 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.923973 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-6xws7" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.924047 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.926268 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-2m68z"] Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.985452 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-combined-ca-bundle\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.985866 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-scripts\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.986043 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-config-data\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:05 crc kubenswrapper[4871]: I1007 23:55:05.986078 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p74j\" (UniqueName: \"kubernetes.io/projected/f7e17c14-ed21-4d89-9299-da86a522d841-kube-api-access-8p74j\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.088626 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-scripts\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.088758 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-config-data\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.088817 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p74j\" (UniqueName: \"kubernetes.io/projected/f7e17c14-ed21-4d89-9299-da86a522d841-kube-api-access-8p74j\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.088901 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-combined-ca-bundle\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.103714 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-scripts\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.104101 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-config-data\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.104697 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-combined-ca-bundle\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.108515 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p74j\" (UniqueName: \"kubernetes.io/projected/f7e17c14-ed21-4d89-9299-da86a522d841-kube-api-access-8p74j\") pod \"aodh-db-sync-2m68z\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.251617 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.335865 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" exitCode=0 Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.335923 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104"} Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.335964 4871 scope.go:117] "RemoveContainer" containerID="cee94e818dbd839f5c342b9e271035e3e73be566c36380410bb63a203783329e" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.336762 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:55:06 crc kubenswrapper[4871]: E1007 23:55:06.337107 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:55:06 crc kubenswrapper[4871]: I1007 23:55:06.841957 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-2m68z"] Oct 07 23:55:06 crc kubenswrapper[4871]: W1007 23:55:06.843900 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7e17c14_ed21_4d89_9299_da86a522d841.slice/crio-babd0ceba1dd164efb625ea714bce024fa847d081556feafc23857e91764fa51 WatchSource:0}: Error finding container babd0ceba1dd164efb625ea714bce024fa847d081556feafc23857e91764fa51: Status 404 returned error can't find the container with id babd0ceba1dd164efb625ea714bce024fa847d081556feafc23857e91764fa51 Oct 07 23:55:07 crc kubenswrapper[4871]: I1007 23:55:07.056348 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-6084-account-create-jcwkq"] Oct 07 23:55:07 crc kubenswrapper[4871]: I1007 23:55:07.064936 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-6084-account-create-jcwkq"] Oct 07 23:55:07 crc kubenswrapper[4871]: I1007 23:55:07.348663 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2m68z" event={"ID":"f7e17c14-ed21-4d89-9299-da86a522d841","Type":"ContainerStarted","Data":"babd0ceba1dd164efb625ea714bce024fa847d081556feafc23857e91764fa51"} Oct 07 23:55:08 crc kubenswrapper[4871]: I1007 23:55:08.996441 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d193011-4f44-4db3-837a-16116f2ab419" path="/var/lib/kubelet/pods/8d193011-4f44-4db3-837a-16116f2ab419/volumes" Oct 07 23:55:11 crc kubenswrapper[4871]: I1007 23:55:11.683667 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 23:55:13 crc kubenswrapper[4871]: I1007 23:55:13.414601 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2m68z" event={"ID":"f7e17c14-ed21-4d89-9299-da86a522d841","Type":"ContainerStarted","Data":"d87db975bcc00bb201a13424f690fb19591106a0d107096d9360f72ea5b5cdb8"} Oct 07 23:55:13 crc kubenswrapper[4871]: I1007 23:55:13.438287 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-2m68z" podStartSLOduration=3.103916149 podStartE2EDuration="8.438261905s" podCreationTimestamp="2025-10-07 23:55:05 +0000 UTC" firstStartedPulling="2025-10-07 23:55:06.850629997 +0000 UTC m=+6380.653328070" lastFinishedPulling="2025-10-07 23:55:12.184975753 +0000 UTC m=+6385.987673826" observedRunningTime="2025-10-07 23:55:13.426399164 +0000 UTC m=+6387.229097247" watchObservedRunningTime="2025-10-07 23:55:13.438261905 +0000 UTC m=+6387.240959978" Oct 07 23:55:14 crc kubenswrapper[4871]: I1007 23:55:14.033132 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-86mch"] Oct 07 23:55:14 crc kubenswrapper[4871]: I1007 23:55:14.045585 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-86mch"] Oct 07 23:55:15 crc kubenswrapper[4871]: I1007 23:55:15.002693 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ede6bf1-ea5c-4298-9c61-5330dfc70805" path="/var/lib/kubelet/pods/3ede6bf1-ea5c-4298-9c61-5330dfc70805/volumes" Oct 07 23:55:15 crc kubenswrapper[4871]: I1007 23:55:15.438257 4871 generic.go:334] "Generic (PLEG): container finished" podID="f7e17c14-ed21-4d89-9299-da86a522d841" containerID="d87db975bcc00bb201a13424f690fb19591106a0d107096d9360f72ea5b5cdb8" exitCode=0 Oct 07 23:55:15 crc kubenswrapper[4871]: I1007 23:55:15.438300 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2m68z" event={"ID":"f7e17c14-ed21-4d89-9299-da86a522d841","Type":"ContainerDied","Data":"d87db975bcc00bb201a13424f690fb19591106a0d107096d9360f72ea5b5cdb8"} Oct 07 23:55:16 crc kubenswrapper[4871]: I1007 23:55:16.851784 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.022706 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-config-data\") pod \"f7e17c14-ed21-4d89-9299-da86a522d841\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.022765 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p74j\" (UniqueName: \"kubernetes.io/projected/f7e17c14-ed21-4d89-9299-da86a522d841-kube-api-access-8p74j\") pod \"f7e17c14-ed21-4d89-9299-da86a522d841\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.022992 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-combined-ca-bundle\") pod \"f7e17c14-ed21-4d89-9299-da86a522d841\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.023425 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-scripts\") pod \"f7e17c14-ed21-4d89-9299-da86a522d841\" (UID: \"f7e17c14-ed21-4d89-9299-da86a522d841\") " Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.028327 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-scripts" (OuterVolumeSpecName: "scripts") pod "f7e17c14-ed21-4d89-9299-da86a522d841" (UID: "f7e17c14-ed21-4d89-9299-da86a522d841"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.030092 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e17c14-ed21-4d89-9299-da86a522d841-kube-api-access-8p74j" (OuterVolumeSpecName: "kube-api-access-8p74j") pod "f7e17c14-ed21-4d89-9299-da86a522d841" (UID: "f7e17c14-ed21-4d89-9299-da86a522d841"). InnerVolumeSpecName "kube-api-access-8p74j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.057495 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-config-data" (OuterVolumeSpecName: "config-data") pod "f7e17c14-ed21-4d89-9299-da86a522d841" (UID: "f7e17c14-ed21-4d89-9299-da86a522d841"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.083983 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7e17c14-ed21-4d89-9299-da86a522d841" (UID: "f7e17c14-ed21-4d89-9299-da86a522d841"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.128043 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.128111 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.128141 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e17c14-ed21-4d89-9299-da86a522d841-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.128167 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p74j\" (UniqueName: \"kubernetes.io/projected/f7e17c14-ed21-4d89-9299-da86a522d841-kube-api-access-8p74j\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.461687 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2m68z" event={"ID":"f7e17c14-ed21-4d89-9299-da86a522d841","Type":"ContainerDied","Data":"babd0ceba1dd164efb625ea714bce024fa847d081556feafc23857e91764fa51"} Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.461740 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="babd0ceba1dd164efb625ea714bce024fa847d081556feafc23857e91764fa51" Oct 07 23:55:17 crc kubenswrapper[4871]: I1007 23:55:17.461867 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2m68z" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.477610 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 07 23:55:20 crc kubenswrapper[4871]: E1007 23:55:20.478649 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e17c14-ed21-4d89-9299-da86a522d841" containerName="aodh-db-sync" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.478666 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e17c14-ed21-4d89-9299-da86a522d841" containerName="aodh-db-sync" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.478931 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e17c14-ed21-4d89-9299-da86a522d841" containerName="aodh-db-sync" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.481290 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.486909 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.487400 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.487647 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-6xws7" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.495443 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.520418 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed1a54c1-15bb-427b-af24-0fb7909985af-scripts\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.520542 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed1a54c1-15bb-427b-af24-0fb7909985af-config-data\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.520572 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h62tz\" (UniqueName: \"kubernetes.io/projected/ed1a54c1-15bb-427b-af24-0fb7909985af-kube-api-access-h62tz\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.520700 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1a54c1-15bb-427b-af24-0fb7909985af-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.622344 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed1a54c1-15bb-427b-af24-0fb7909985af-scripts\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.622418 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed1a54c1-15bb-427b-af24-0fb7909985af-config-data\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.622447 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h62tz\" (UniqueName: \"kubernetes.io/projected/ed1a54c1-15bb-427b-af24-0fb7909985af-kube-api-access-h62tz\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.623880 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1a54c1-15bb-427b-af24-0fb7909985af-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.629754 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed1a54c1-15bb-427b-af24-0fb7909985af-scripts\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.634353 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed1a54c1-15bb-427b-af24-0fb7909985af-config-data\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.639124 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1a54c1-15bb-427b-af24-0fb7909985af-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.643063 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h62tz\" (UniqueName: \"kubernetes.io/projected/ed1a54c1-15bb-427b-af24-0fb7909985af-kube-api-access-h62tz\") pod \"aodh-0\" (UID: \"ed1a54c1-15bb-427b-af24-0fb7909985af\") " pod="openstack/aodh-0" Oct 07 23:55:20 crc kubenswrapper[4871]: I1007 23:55:20.818312 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 07 23:55:21 crc kubenswrapper[4871]: I1007 23:55:21.335929 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 07 23:55:21 crc kubenswrapper[4871]: I1007 23:55:21.516237 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ed1a54c1-15bb-427b-af24-0fb7909985af","Type":"ContainerStarted","Data":"09ac49035f51ce5416bf73a2451762390024f403b1478d7e5b66fc66f05e2df8"} Oct 07 23:55:21 crc kubenswrapper[4871]: I1007 23:55:21.982497 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:55:21 crc kubenswrapper[4871]: E1007 23:55:21.983096 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:55:22 crc kubenswrapper[4871]: I1007 23:55:22.524890 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ed1a54c1-15bb-427b-af24-0fb7909985af","Type":"ContainerStarted","Data":"ee702fa7d10c8732cdad5a58de17ceed05bc3f8dcef8cdea4fd170fbf846bee0"} Oct 07 23:55:23 crc kubenswrapper[4871]: I1007 23:55:23.257069 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:55:23 crc kubenswrapper[4871]: I1007 23:55:23.257600 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="ceilometer-central-agent" containerID="cri-o://90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea" gracePeriod=30 Oct 07 23:55:23 crc kubenswrapper[4871]: I1007 23:55:23.258064 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="proxy-httpd" containerID="cri-o://99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193" gracePeriod=30 Oct 07 23:55:23 crc kubenswrapper[4871]: I1007 23:55:23.258114 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="sg-core" containerID="cri-o://2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21" gracePeriod=30 Oct 07 23:55:23 crc kubenswrapper[4871]: I1007 23:55:23.258144 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="ceilometer-notification-agent" containerID="cri-o://7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b" gracePeriod=30 Oct 07 23:55:23 crc kubenswrapper[4871]: I1007 23:55:23.542439 4871 generic.go:334] "Generic (PLEG): container finished" podID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerID="99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193" exitCode=0 Oct 07 23:55:23 crc kubenswrapper[4871]: I1007 23:55:23.542470 4871 generic.go:334] "Generic (PLEG): container finished" podID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerID="2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21" exitCode=2 Oct 07 23:55:23 crc kubenswrapper[4871]: I1007 23:55:23.542493 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerDied","Data":"99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193"} Oct 07 23:55:23 crc kubenswrapper[4871]: I1007 23:55:23.542517 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerDied","Data":"2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21"} Oct 07 23:55:24 crc kubenswrapper[4871]: I1007 23:55:24.556883 4871 generic.go:334] "Generic (PLEG): container finished" podID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerID="90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea" exitCode=0 Oct 07 23:55:24 crc kubenswrapper[4871]: I1007 23:55:24.556931 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerDied","Data":"90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea"} Oct 07 23:55:24 crc kubenswrapper[4871]: I1007 23:55:24.560786 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ed1a54c1-15bb-427b-af24-0fb7909985af","Type":"ContainerStarted","Data":"e950dac20cbe208972fdd6d90b3236ddfa1508bba63175491eb0264dbc436489"} Oct 07 23:55:25 crc kubenswrapper[4871]: I1007 23:55:25.570935 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ed1a54c1-15bb-427b-af24-0fb7909985af","Type":"ContainerStarted","Data":"d2eb3dfb6dafefe54e4dcc05553c9acb7c41508b2da3a2d9b3338b5b6ca7212e"} Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.397011 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.472105 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-run-httpd\") pod \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.472365 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-config-data\") pod \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.472483 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-combined-ca-bundle\") pod \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.472588 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-scripts\") pod \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.472762 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbdr8\" (UniqueName: \"kubernetes.io/projected/0a81fdaf-895c-415c-b921-6bd563b3eeb5-kube-api-access-gbdr8\") pod \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.472523 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0a81fdaf-895c-415c-b921-6bd563b3eeb5" (UID: "0a81fdaf-895c-415c-b921-6bd563b3eeb5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.473010 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-log-httpd\") pod \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.473124 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-sg-core-conf-yaml\") pod \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\" (UID: \"0a81fdaf-895c-415c-b921-6bd563b3eeb5\") " Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.473490 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0a81fdaf-895c-415c-b921-6bd563b3eeb5" (UID: "0a81fdaf-895c-415c-b921-6bd563b3eeb5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.474034 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.474104 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a81fdaf-895c-415c-b921-6bd563b3eeb5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.489993 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a81fdaf-895c-415c-b921-6bd563b3eeb5-kube-api-access-gbdr8" (OuterVolumeSpecName: "kube-api-access-gbdr8") pod "0a81fdaf-895c-415c-b921-6bd563b3eeb5" (UID: "0a81fdaf-895c-415c-b921-6bd563b3eeb5"). InnerVolumeSpecName "kube-api-access-gbdr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.490717 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-scripts" (OuterVolumeSpecName: "scripts") pod "0a81fdaf-895c-415c-b921-6bd563b3eeb5" (UID: "0a81fdaf-895c-415c-b921-6bd563b3eeb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.515352 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0a81fdaf-895c-415c-b921-6bd563b3eeb5" (UID: "0a81fdaf-895c-415c-b921-6bd563b3eeb5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.565840 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a81fdaf-895c-415c-b921-6bd563b3eeb5" (UID: "0a81fdaf-895c-415c-b921-6bd563b3eeb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.575643 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.575671 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.575681 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbdr8\" (UniqueName: \"kubernetes.io/projected/0a81fdaf-895c-415c-b921-6bd563b3eeb5-kube-api-access-gbdr8\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.575691 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.592135 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-config-data" (OuterVolumeSpecName: "config-data") pod "0a81fdaf-895c-415c-b921-6bd563b3eeb5" (UID: "0a81fdaf-895c-415c-b921-6bd563b3eeb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.603701 4871 generic.go:334] "Generic (PLEG): container finished" podID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerID="7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b" exitCode=0 Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.603905 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.603914 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerDied","Data":"7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b"} Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.604589 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a81fdaf-895c-415c-b921-6bd563b3eeb5","Type":"ContainerDied","Data":"ab9b0dd0b7bf8eb4288cd15b518235ae0ce02048ac711596e3bbf55080ed09e4"} Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.604618 4871 scope.go:117] "RemoveContainer" containerID="99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.607388 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ed1a54c1-15bb-427b-af24-0fb7909985af","Type":"ContainerStarted","Data":"ccd1c3070310757ecc1aae39823df5ee04ad58126be1b403b6af87ee1bef1374"} Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.642241 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.4555504409999998 podStartE2EDuration="7.642220564s" podCreationTimestamp="2025-10-07 23:55:20 +0000 UTC" firstStartedPulling="2025-10-07 23:55:21.353506746 +0000 UTC m=+6395.156204979" lastFinishedPulling="2025-10-07 23:55:26.540177029 +0000 UTC m=+6400.342875102" observedRunningTime="2025-10-07 23:55:27.632860028 +0000 UTC m=+6401.435558101" watchObservedRunningTime="2025-10-07 23:55:27.642220564 +0000 UTC m=+6401.444918637" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.683362 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a81fdaf-895c-415c-b921-6bd563b3eeb5-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.691484 4871 scope.go:117] "RemoveContainer" containerID="2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.701004 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.721999 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.723988 4871 scope.go:117] "RemoveContainer" containerID="7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.735764 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:55:27 crc kubenswrapper[4871]: E1007 23:55:27.736318 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="ceilometer-notification-agent" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.736336 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="ceilometer-notification-agent" Oct 07 23:55:27 crc kubenswrapper[4871]: E1007 23:55:27.736355 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="ceilometer-central-agent" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.736362 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="ceilometer-central-agent" Oct 07 23:55:27 crc kubenswrapper[4871]: E1007 23:55:27.736407 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="sg-core" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.736413 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="sg-core" Oct 07 23:55:27 crc kubenswrapper[4871]: E1007 23:55:27.736426 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="proxy-httpd" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.736431 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="proxy-httpd" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.736607 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="proxy-httpd" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.736625 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="ceilometer-notification-agent" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.736642 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="sg-core" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.736651 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" containerName="ceilometer-central-agent" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.738490 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.742256 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.742491 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.745666 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.750710 4871 scope.go:117] "RemoveContainer" containerID="90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.772698 4871 scope.go:117] "RemoveContainer" containerID="99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193" Oct 07 23:55:27 crc kubenswrapper[4871]: E1007 23:55:27.773175 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193\": container with ID starting with 99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193 not found: ID does not exist" containerID="99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.773206 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193"} err="failed to get container status \"99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193\": rpc error: code = NotFound desc = could not find container \"99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193\": container with ID starting with 99dc54ec8821e53f25f98aee480197ad48436c6e75095f6df0a8e24a45eb2193 not found: ID does not exist" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.773226 4871 scope.go:117] "RemoveContainer" containerID="2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21" Oct 07 23:55:27 crc kubenswrapper[4871]: E1007 23:55:27.774873 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21\": container with ID starting with 2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21 not found: ID does not exist" containerID="2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.774898 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21"} err="failed to get container status \"2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21\": rpc error: code = NotFound desc = could not find container \"2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21\": container with ID starting with 2679b452a9e06f2a70c75a496ea87c1227ef7f8917f127dd06399f98ff7a1b21 not found: ID does not exist" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.774911 4871 scope.go:117] "RemoveContainer" containerID="7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b" Oct 07 23:55:27 crc kubenswrapper[4871]: E1007 23:55:27.775169 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b\": container with ID starting with 7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b not found: ID does not exist" containerID="7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.775189 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b"} err="failed to get container status \"7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b\": rpc error: code = NotFound desc = could not find container \"7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b\": container with ID starting with 7750bf61a598c438f5af29624d414bc38825bb78d24927ec6a2495cf0618f75b not found: ID does not exist" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.775202 4871 scope.go:117] "RemoveContainer" containerID="90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea" Oct 07 23:55:27 crc kubenswrapper[4871]: E1007 23:55:27.775974 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea\": container with ID starting with 90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea not found: ID does not exist" containerID="90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.776016 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea"} err="failed to get container status \"90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea\": rpc error: code = NotFound desc = could not find container \"90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea\": container with ID starting with 90823fedc3bb358809ec1af2e7f55abf0ddeb48e1e8edaec73480df4774246ea not found: ID does not exist" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.887224 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc4vx\" (UniqueName: \"kubernetes.io/projected/2f03be08-f24c-4da2-9238-191cf91ed53e-kube-api-access-wc4vx\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.887587 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.887828 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-run-httpd\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.888044 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-log-httpd\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.888158 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-config-data\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.888214 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.888437 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-scripts\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.990475 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-run-httpd\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.990571 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-log-httpd\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.990628 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-config-data\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.990655 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.990749 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-scripts\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.990829 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc4vx\" (UniqueName: \"kubernetes.io/projected/2f03be08-f24c-4da2-9238-191cf91ed53e-kube-api-access-wc4vx\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.990910 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.991525 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-log-httpd\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:27 crc kubenswrapper[4871]: I1007 23:55:27.991865 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-run-httpd\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:28 crc kubenswrapper[4871]: I1007 23:55:28.003251 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-scripts\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:28 crc kubenswrapper[4871]: I1007 23:55:28.008613 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:28 crc kubenswrapper[4871]: I1007 23:55:28.011180 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:28 crc kubenswrapper[4871]: I1007 23:55:28.011232 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-config-data\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:28 crc kubenswrapper[4871]: I1007 23:55:28.028365 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc4vx\" (UniqueName: \"kubernetes.io/projected/2f03be08-f24c-4da2-9238-191cf91ed53e-kube-api-access-wc4vx\") pod \"ceilometer-0\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " pod="openstack/ceilometer-0" Oct 07 23:55:28 crc kubenswrapper[4871]: I1007 23:55:28.061244 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:55:28 crc kubenswrapper[4871]: I1007 23:55:28.581715 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:55:28 crc kubenswrapper[4871]: W1007 23:55:28.589742 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f03be08_f24c_4da2_9238_191cf91ed53e.slice/crio-04c568746fcc5946ccd6da8f62bf7fa4c1f17ac083c60ee29ec2396915cab846 WatchSource:0}: Error finding container 04c568746fcc5946ccd6da8f62bf7fa4c1f17ac083c60ee29ec2396915cab846: Status 404 returned error can't find the container with id 04c568746fcc5946ccd6da8f62bf7fa4c1f17ac083c60ee29ec2396915cab846 Oct 07 23:55:28 crc kubenswrapper[4871]: I1007 23:55:28.622964 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerStarted","Data":"04c568746fcc5946ccd6da8f62bf7fa4c1f17ac083c60ee29ec2396915cab846"} Oct 07 23:55:28 crc kubenswrapper[4871]: I1007 23:55:28.994806 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a81fdaf-895c-415c-b921-6bd563b3eeb5" path="/var/lib/kubelet/pods/0a81fdaf-895c-415c-b921-6bd563b3eeb5/volumes" Oct 07 23:55:29 crc kubenswrapper[4871]: I1007 23:55:29.632594 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerStarted","Data":"780c4a6b26a5c4183e9a1858ae12d291a47de42e792060b23c0c4c6101a75c91"} Oct 07 23:55:30 crc kubenswrapper[4871]: I1007 23:55:30.649102 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerStarted","Data":"7161a6179779a43091b0473bfee661ee11ee7a917eaace0d1d7008ce37e83fe8"} Oct 07 23:55:30 crc kubenswrapper[4871]: I1007 23:55:30.649578 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerStarted","Data":"d4a709874a19f6598ca7556edc2929abf9f7375e8d7fbdd47e85dc3fe232a8b9"} Oct 07 23:55:32 crc kubenswrapper[4871]: I1007 23:55:32.682428 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerStarted","Data":"e877c05868f81f1a524cc43a685ca654e4ff009aa5d3785794b1f2b5b908f7ac"} Oct 07 23:55:32 crc kubenswrapper[4871]: I1007 23:55:32.683185 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 23:55:32 crc kubenswrapper[4871]: I1007 23:55:32.706302 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.592733394 podStartE2EDuration="5.70628467s" podCreationTimestamp="2025-10-07 23:55:27 +0000 UTC" firstStartedPulling="2025-10-07 23:55:28.592014316 +0000 UTC m=+6402.394712389" lastFinishedPulling="2025-10-07 23:55:31.705565602 +0000 UTC m=+6405.508263665" observedRunningTime="2025-10-07 23:55:32.704911144 +0000 UTC m=+6406.507609207" watchObservedRunningTime="2025-10-07 23:55:32.70628467 +0000 UTC m=+6406.508982743" Oct 07 23:55:33 crc kubenswrapper[4871]: I1007 23:55:33.371091 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-mgfvv"] Oct 07 23:55:33 crc kubenswrapper[4871]: I1007 23:55:33.372785 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mgfvv" Oct 07 23:55:33 crc kubenswrapper[4871]: I1007 23:55:33.379184 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mgfvv"] Oct 07 23:55:33 crc kubenswrapper[4871]: I1007 23:55:33.412000 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2glq\" (UniqueName: \"kubernetes.io/projected/24f2709b-5908-4273-a24b-53af7ff871da-kube-api-access-m2glq\") pod \"manila-db-create-mgfvv\" (UID: \"24f2709b-5908-4273-a24b-53af7ff871da\") " pod="openstack/manila-db-create-mgfvv" Oct 07 23:55:33 crc kubenswrapper[4871]: I1007 23:55:33.514083 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2glq\" (UniqueName: \"kubernetes.io/projected/24f2709b-5908-4273-a24b-53af7ff871da-kube-api-access-m2glq\") pod \"manila-db-create-mgfvv\" (UID: \"24f2709b-5908-4273-a24b-53af7ff871da\") " pod="openstack/manila-db-create-mgfvv" Oct 07 23:55:33 crc kubenswrapper[4871]: I1007 23:55:33.532952 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2glq\" (UniqueName: \"kubernetes.io/projected/24f2709b-5908-4273-a24b-53af7ff871da-kube-api-access-m2glq\") pod \"manila-db-create-mgfvv\" (UID: \"24f2709b-5908-4273-a24b-53af7ff871da\") " pod="openstack/manila-db-create-mgfvv" Oct 07 23:55:33 crc kubenswrapper[4871]: I1007 23:55:33.694865 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mgfvv" Oct 07 23:55:33 crc kubenswrapper[4871]: I1007 23:55:33.983726 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:55:33 crc kubenswrapper[4871]: E1007 23:55:33.984354 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:55:34 crc kubenswrapper[4871]: I1007 23:55:34.357899 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mgfvv"] Oct 07 23:55:34 crc kubenswrapper[4871]: W1007 23:55:34.359911 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24f2709b_5908_4273_a24b_53af7ff871da.slice/crio-09a3de2eb1a0aced95553a0cb3b1011cf504d96b82dc50e3a773df0bbd1204fa WatchSource:0}: Error finding container 09a3de2eb1a0aced95553a0cb3b1011cf504d96b82dc50e3a773df0bbd1204fa: Status 404 returned error can't find the container with id 09a3de2eb1a0aced95553a0cb3b1011cf504d96b82dc50e3a773df0bbd1204fa Oct 07 23:55:34 crc kubenswrapper[4871]: I1007 23:55:34.712724 4871 generic.go:334] "Generic (PLEG): container finished" podID="24f2709b-5908-4273-a24b-53af7ff871da" containerID="9051282a100a0f323719745bea42ef9a23c3813f748ca05f66e0f27884a1a792" exitCode=0 Oct 07 23:55:34 crc kubenswrapper[4871]: I1007 23:55:34.712824 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mgfvv" event={"ID":"24f2709b-5908-4273-a24b-53af7ff871da","Type":"ContainerDied","Data":"9051282a100a0f323719745bea42ef9a23c3813f748ca05f66e0f27884a1a792"} Oct 07 23:55:34 crc kubenswrapper[4871]: I1007 23:55:34.713060 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mgfvv" event={"ID":"24f2709b-5908-4273-a24b-53af7ff871da","Type":"ContainerStarted","Data":"09a3de2eb1a0aced95553a0cb3b1011cf504d96b82dc50e3a773df0bbd1204fa"} Oct 07 23:55:36 crc kubenswrapper[4871]: I1007 23:55:36.178196 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mgfvv" Oct 07 23:55:36 crc kubenswrapper[4871]: I1007 23:55:36.294297 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2glq\" (UniqueName: \"kubernetes.io/projected/24f2709b-5908-4273-a24b-53af7ff871da-kube-api-access-m2glq\") pod \"24f2709b-5908-4273-a24b-53af7ff871da\" (UID: \"24f2709b-5908-4273-a24b-53af7ff871da\") " Oct 07 23:55:36 crc kubenswrapper[4871]: I1007 23:55:36.300707 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f2709b-5908-4273-a24b-53af7ff871da-kube-api-access-m2glq" (OuterVolumeSpecName: "kube-api-access-m2glq") pod "24f2709b-5908-4273-a24b-53af7ff871da" (UID: "24f2709b-5908-4273-a24b-53af7ff871da"). InnerVolumeSpecName "kube-api-access-m2glq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:55:36 crc kubenswrapper[4871]: I1007 23:55:36.396899 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2glq\" (UniqueName: \"kubernetes.io/projected/24f2709b-5908-4273-a24b-53af7ff871da-kube-api-access-m2glq\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:36 crc kubenswrapper[4871]: I1007 23:55:36.755994 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mgfvv" event={"ID":"24f2709b-5908-4273-a24b-53af7ff871da","Type":"ContainerDied","Data":"09a3de2eb1a0aced95553a0cb3b1011cf504d96b82dc50e3a773df0bbd1204fa"} Oct 07 23:55:36 crc kubenswrapper[4871]: I1007 23:55:36.756036 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09a3de2eb1a0aced95553a0cb3b1011cf504d96b82dc50e3a773df0bbd1204fa" Oct 07 23:55:36 crc kubenswrapper[4871]: I1007 23:55:36.756133 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mgfvv" Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.491124 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-721c-account-create-nxz6g"] Oct 07 23:55:43 crc kubenswrapper[4871]: E1007 23:55:43.493010 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f2709b-5908-4273-a24b-53af7ff871da" containerName="mariadb-database-create" Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.493123 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f2709b-5908-4273-a24b-53af7ff871da" containerName="mariadb-database-create" Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.493398 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f2709b-5908-4273-a24b-53af7ff871da" containerName="mariadb-database-create" Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.494242 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-721c-account-create-nxz6g" Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.496700 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.509247 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-721c-account-create-nxz6g"] Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.565672 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5s6h\" (UniqueName: \"kubernetes.io/projected/54c505df-ac87-41ee-95cf-7e23638f1973-kube-api-access-r5s6h\") pod \"manila-721c-account-create-nxz6g\" (UID: \"54c505df-ac87-41ee-95cf-7e23638f1973\") " pod="openstack/manila-721c-account-create-nxz6g" Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.667639 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5s6h\" (UniqueName: \"kubernetes.io/projected/54c505df-ac87-41ee-95cf-7e23638f1973-kube-api-access-r5s6h\") pod \"manila-721c-account-create-nxz6g\" (UID: \"54c505df-ac87-41ee-95cf-7e23638f1973\") " pod="openstack/manila-721c-account-create-nxz6g" Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.683728 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5s6h\" (UniqueName: \"kubernetes.io/projected/54c505df-ac87-41ee-95cf-7e23638f1973-kube-api-access-r5s6h\") pod \"manila-721c-account-create-nxz6g\" (UID: \"54c505df-ac87-41ee-95cf-7e23638f1973\") " pod="openstack/manila-721c-account-create-nxz6g" Oct 07 23:55:43 crc kubenswrapper[4871]: I1007 23:55:43.826849 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-721c-account-create-nxz6g" Oct 07 23:55:44 crc kubenswrapper[4871]: I1007 23:55:44.330253 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-721c-account-create-nxz6g"] Oct 07 23:55:44 crc kubenswrapper[4871]: W1007 23:55:44.339617 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54c505df_ac87_41ee_95cf_7e23638f1973.slice/crio-8e560fd9339bfa23c621951e7fb5cb5f9eedb79f144a4d82d9ff15a358bf2a2d WatchSource:0}: Error finding container 8e560fd9339bfa23c621951e7fb5cb5f9eedb79f144a4d82d9ff15a358bf2a2d: Status 404 returned error can't find the container with id 8e560fd9339bfa23c621951e7fb5cb5f9eedb79f144a4d82d9ff15a358bf2a2d Oct 07 23:55:44 crc kubenswrapper[4871]: I1007 23:55:44.850109 4871 generic.go:334] "Generic (PLEG): container finished" podID="54c505df-ac87-41ee-95cf-7e23638f1973" containerID="4d3692f0b1d79ab21fff7720b1014af96a533223d1e5841bb6c0564d6708f640" exitCode=0 Oct 07 23:55:44 crc kubenswrapper[4871]: I1007 23:55:44.850193 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-721c-account-create-nxz6g" event={"ID":"54c505df-ac87-41ee-95cf-7e23638f1973","Type":"ContainerDied","Data":"4d3692f0b1d79ab21fff7720b1014af96a533223d1e5841bb6c0564d6708f640"} Oct 07 23:55:44 crc kubenswrapper[4871]: I1007 23:55:44.850411 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-721c-account-create-nxz6g" event={"ID":"54c505df-ac87-41ee-95cf-7e23638f1973","Type":"ContainerStarted","Data":"8e560fd9339bfa23c621951e7fb5cb5f9eedb79f144a4d82d9ff15a358bf2a2d"} Oct 07 23:55:46 crc kubenswrapper[4871]: I1007 23:55:46.303265 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-721c-account-create-nxz6g" Oct 07 23:55:46 crc kubenswrapper[4871]: I1007 23:55:46.432986 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5s6h\" (UniqueName: \"kubernetes.io/projected/54c505df-ac87-41ee-95cf-7e23638f1973-kube-api-access-r5s6h\") pod \"54c505df-ac87-41ee-95cf-7e23638f1973\" (UID: \"54c505df-ac87-41ee-95cf-7e23638f1973\") " Oct 07 23:55:46 crc kubenswrapper[4871]: I1007 23:55:46.441246 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c505df-ac87-41ee-95cf-7e23638f1973-kube-api-access-r5s6h" (OuterVolumeSpecName: "kube-api-access-r5s6h") pod "54c505df-ac87-41ee-95cf-7e23638f1973" (UID: "54c505df-ac87-41ee-95cf-7e23638f1973"). InnerVolumeSpecName "kube-api-access-r5s6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:55:46 crc kubenswrapper[4871]: I1007 23:55:46.536019 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5s6h\" (UniqueName: \"kubernetes.io/projected/54c505df-ac87-41ee-95cf-7e23638f1973-kube-api-access-r5s6h\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:46 crc kubenswrapper[4871]: I1007 23:55:46.894195 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-721c-account-create-nxz6g" event={"ID":"54c505df-ac87-41ee-95cf-7e23638f1973","Type":"ContainerDied","Data":"8e560fd9339bfa23c621951e7fb5cb5f9eedb79f144a4d82d9ff15a358bf2a2d"} Oct 07 23:55:46 crc kubenswrapper[4871]: I1007 23:55:46.894250 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e560fd9339bfa23c621951e7fb5cb5f9eedb79f144a4d82d9ff15a358bf2a2d" Oct 07 23:55:46 crc kubenswrapper[4871]: I1007 23:55:46.894317 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-721c-account-create-nxz6g" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.864815 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-8qtck"] Oct 07 23:55:48 crc kubenswrapper[4871]: E1007 23:55:48.865906 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c505df-ac87-41ee-95cf-7e23638f1973" containerName="mariadb-account-create" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.865926 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c505df-ac87-41ee-95cf-7e23638f1973" containerName="mariadb-account-create" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.866227 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c505df-ac87-41ee-95cf-7e23638f1973" containerName="mariadb-account-create" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.867117 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.870679 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-n4m47" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.884756 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.887190 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-job-config-data\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.887341 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jff4\" (UniqueName: \"kubernetes.io/projected/e18ea361-7369-4dca-827a-74d36b75bf7d-kube-api-access-5jff4\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.887394 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-config-data\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.887412 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-combined-ca-bundle\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.892784 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-8qtck"] Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.982271 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:55:48 crc kubenswrapper[4871]: E1007 23:55:48.982603 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.989194 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-job-config-data\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.989332 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jff4\" (UniqueName: \"kubernetes.io/projected/e18ea361-7369-4dca-827a-74d36b75bf7d-kube-api-access-5jff4\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.989395 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-config-data\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.989413 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-combined-ca-bundle\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.995561 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-combined-ca-bundle\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:48 crc kubenswrapper[4871]: I1007 23:55:48.997667 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-config-data\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:49 crc kubenswrapper[4871]: I1007 23:55:49.002145 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-job-config-data\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:49 crc kubenswrapper[4871]: I1007 23:55:49.015141 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jff4\" (UniqueName: \"kubernetes.io/projected/e18ea361-7369-4dca-827a-74d36b75bf7d-kube-api-access-5jff4\") pod \"manila-db-sync-8qtck\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:49 crc kubenswrapper[4871]: I1007 23:55:49.212545 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:49 crc kubenswrapper[4871]: I1007 23:55:49.949081 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-8qtck"] Oct 07 23:55:50 crc kubenswrapper[4871]: I1007 23:55:50.948010 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-8qtck" event={"ID":"e18ea361-7369-4dca-827a-74d36b75bf7d","Type":"ContainerStarted","Data":"594b0db64ba2eb6d74c914dd2512a57a4859a063988c9471bb67ff759af52d66"} Oct 07 23:55:54 crc kubenswrapper[4871]: I1007 23:55:54.998310 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-8qtck" event={"ID":"e18ea361-7369-4dca-827a-74d36b75bf7d","Type":"ContainerStarted","Data":"5e14d41415c1fb260632ce64ea841dfbe5acd9a67e2e210aa508333a06753076"} Oct 07 23:55:55 crc kubenswrapper[4871]: I1007 23:55:55.024032 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-8qtck" podStartSLOduration=2.585352824 podStartE2EDuration="7.024014227s" podCreationTimestamp="2025-10-07 23:55:48 +0000 UTC" firstStartedPulling="2025-10-07 23:55:49.952625508 +0000 UTC m=+6423.755323581" lastFinishedPulling="2025-10-07 23:55:54.391286911 +0000 UTC m=+6428.193984984" observedRunningTime="2025-10-07 23:55:55.011209921 +0000 UTC m=+6428.813907994" watchObservedRunningTime="2025-10-07 23:55:55.024014227 +0000 UTC m=+6428.826712300" Oct 07 23:55:57 crc kubenswrapper[4871]: I1007 23:55:57.061815 4871 generic.go:334] "Generic (PLEG): container finished" podID="e18ea361-7369-4dca-827a-74d36b75bf7d" containerID="5e14d41415c1fb260632ce64ea841dfbe5acd9a67e2e210aa508333a06753076" exitCode=0 Oct 07 23:55:57 crc kubenswrapper[4871]: I1007 23:55:57.062388 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-8qtck" event={"ID":"e18ea361-7369-4dca-827a-74d36b75bf7d","Type":"ContainerDied","Data":"5e14d41415c1fb260632ce64ea841dfbe5acd9a67e2e210aa508333a06753076"} Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.072077 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.688849 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.804189 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jff4\" (UniqueName: \"kubernetes.io/projected/e18ea361-7369-4dca-827a-74d36b75bf7d-kube-api-access-5jff4\") pod \"e18ea361-7369-4dca-827a-74d36b75bf7d\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.804352 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-config-data\") pod \"e18ea361-7369-4dca-827a-74d36b75bf7d\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.804399 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-job-config-data\") pod \"e18ea361-7369-4dca-827a-74d36b75bf7d\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.804487 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-combined-ca-bundle\") pod \"e18ea361-7369-4dca-827a-74d36b75bf7d\" (UID: \"e18ea361-7369-4dca-827a-74d36b75bf7d\") " Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.818139 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "e18ea361-7369-4dca-827a-74d36b75bf7d" (UID: "e18ea361-7369-4dca-827a-74d36b75bf7d"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.818321 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e18ea361-7369-4dca-827a-74d36b75bf7d-kube-api-access-5jff4" (OuterVolumeSpecName: "kube-api-access-5jff4") pod "e18ea361-7369-4dca-827a-74d36b75bf7d" (UID: "e18ea361-7369-4dca-827a-74d36b75bf7d"). InnerVolumeSpecName "kube-api-access-5jff4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.821428 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-config-data" (OuterVolumeSpecName: "config-data") pod "e18ea361-7369-4dca-827a-74d36b75bf7d" (UID: "e18ea361-7369-4dca-827a-74d36b75bf7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.841510 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e18ea361-7369-4dca-827a-74d36b75bf7d" (UID: "e18ea361-7369-4dca-827a-74d36b75bf7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.907903 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jff4\" (UniqueName: \"kubernetes.io/projected/e18ea361-7369-4dca-827a-74d36b75bf7d-kube-api-access-5jff4\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.907941 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.907955 4871 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:58 crc kubenswrapper[4871]: I1007 23:55:58.907967 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e18ea361-7369-4dca-827a-74d36b75bf7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.089945 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-8qtck" event={"ID":"e18ea361-7369-4dca-827a-74d36b75bf7d","Type":"ContainerDied","Data":"594b0db64ba2eb6d74c914dd2512a57a4859a063988c9471bb67ff759af52d66"} Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.089998 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="594b0db64ba2eb6d74c914dd2512a57a4859a063988c9471bb67ff759af52d66" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.090047 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-8qtck" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.463498 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 23:55:59 crc kubenswrapper[4871]: E1007 23:55:59.464387 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18ea361-7369-4dca-827a-74d36b75bf7d" containerName="manila-db-sync" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.464411 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18ea361-7369-4dca-827a-74d36b75bf7d" containerName="manila-db-sync" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.464707 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e18ea361-7369-4dca-827a-74d36b75bf7d" containerName="manila-db-sync" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.466088 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.467674 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-n4m47" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.467984 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.468181 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.474196 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.475293 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.478032 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.482226 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.495673 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.505499 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.552180 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f8984b94c-hdzvp"] Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.553880 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.597449 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f8984b94c-hdzvp"] Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622195 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/44e3376b-cce2-4e65-a494-c17253d07c09-ceph\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622244 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/44e3376b-cce2-4e65-a494-c17253d07c09-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622334 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6pql\" (UniqueName: \"kubernetes.io/projected/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-kube-api-access-n6pql\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622358 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-config-data\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622374 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622399 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622441 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q7nj\" (UniqueName: \"kubernetes.io/projected/44e3376b-cce2-4e65-a494-c17253d07c09-kube-api-access-7q7nj\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622483 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-scripts\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622506 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622527 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/44e3376b-cce2-4e65-a494-c17253d07c09-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622546 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-scripts\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622565 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622589 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-config-data\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.622615 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724180 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724238 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-config-data\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724269 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724297 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-config\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724319 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/44e3376b-cce2-4e65-a494-c17253d07c09-ceph\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724339 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/44e3376b-cce2-4e65-a494-c17253d07c09-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724397 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxrdl\" (UniqueName: \"kubernetes.io/projected/65a57c8a-e487-4924-bcfe-8fe57300dc57-kube-api-access-bxrdl\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724414 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-dns-svc\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724448 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6pql\" (UniqueName: \"kubernetes.io/projected/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-kube-api-access-n6pql\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724469 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-config-data\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724485 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-sb\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724502 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724529 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724563 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q7nj\" (UniqueName: \"kubernetes.io/projected/44e3376b-cce2-4e65-a494-c17253d07c09-kube-api-access-7q7nj\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724603 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-scripts\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724619 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-nb\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724638 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724658 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/44e3376b-cce2-4e65-a494-c17253d07c09-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.724678 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-scripts\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.733312 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/44e3376b-cce2-4e65-a494-c17253d07c09-ceph\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.733607 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.733782 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.734834 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.734920 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.735021 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/44e3376b-cce2-4e65-a494-c17253d07c09-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.736054 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.736999 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-scripts\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.737090 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/44e3376b-cce2-4e65-a494-c17253d07c09-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.739966 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.741098 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-config-data\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.743405 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.746818 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44e3376b-cce2-4e65-a494-c17253d07c09-scripts\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.747330 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-config-data\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.750704 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.754487 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q7nj\" (UniqueName: \"kubernetes.io/projected/44e3376b-cce2-4e65-a494-c17253d07c09-kube-api-access-7q7nj\") pod \"manila-share-share1-0\" (UID: \"44e3376b-cce2-4e65-a494-c17253d07c09\") " pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.754763 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6pql\" (UniqueName: \"kubernetes.io/projected/e21ff4e8-d971-4e36-bdb5-ea2e5364326e-kube-api-access-n6pql\") pod \"manila-scheduler-0\" (UID: \"e21ff4e8-d971-4e36-bdb5-ea2e5364326e\") " pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.756077 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.801435 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.806076 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.827583 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-config\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.827847 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-config-data-custom\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.827939 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.828033 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95791187-6a85-4955-997d-920afa9d8adf-logs\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.828105 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxrdl\" (UniqueName: \"kubernetes.io/projected/65a57c8a-e487-4924-bcfe-8fe57300dc57-kube-api-access-bxrdl\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.828175 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-dns-svc\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.828242 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-scripts\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.828323 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-sb\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.828412 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-config-data\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.828492 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjxnr\" (UniqueName: \"kubernetes.io/projected/95791187-6a85-4955-997d-920afa9d8adf-kube-api-access-fjxnr\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.828585 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-nb\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.828654 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95791187-6a85-4955-997d-920afa9d8adf-etc-machine-id\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.829513 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-config\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.831567 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-nb\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.832361 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-dns-svc\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.836784 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-sb\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.850757 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxrdl\" (UniqueName: \"kubernetes.io/projected/65a57c8a-e487-4924-bcfe-8fe57300dc57-kube-api-access-bxrdl\") pod \"dnsmasq-dns-f8984b94c-hdzvp\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.873510 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.930999 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95791187-6a85-4955-997d-920afa9d8adf-logs\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.931057 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-scripts\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.931116 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-config-data\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.931150 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjxnr\" (UniqueName: \"kubernetes.io/projected/95791187-6a85-4955-997d-920afa9d8adf-kube-api-access-fjxnr\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.931203 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95791187-6a85-4955-997d-920afa9d8adf-etc-machine-id\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.931458 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-config-data-custom\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.931482 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.931832 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95791187-6a85-4955-997d-920afa9d8adf-logs\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.935149 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95791187-6a85-4955-997d-920afa9d8adf-etc-machine-id\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.939408 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-config-data-custom\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.944122 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-scripts\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.947369 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.973539 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjxnr\" (UniqueName: \"kubernetes.io/projected/95791187-6a85-4955-997d-920afa9d8adf-kube-api-access-fjxnr\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:55:59 crc kubenswrapper[4871]: I1007 23:55:59.974194 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95791187-6a85-4955-997d-920afa9d8adf-config-data\") pod \"manila-api-0\" (UID: \"95791187-6a85-4955-997d-920afa9d8adf\") " pod="openstack/manila-api-0" Oct 07 23:56:00 crc kubenswrapper[4871]: I1007 23:56:00.048842 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 07 23:56:00 crc kubenswrapper[4871]: I1007 23:56:00.429238 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 23:56:00 crc kubenswrapper[4871]: I1007 23:56:00.548057 4871 scope.go:117] "RemoveContainer" containerID="c3c1dc460812303077ce228afc16b680d56f90a68c150c03cc56a6b658a0096a" Oct 07 23:56:00 crc kubenswrapper[4871]: I1007 23:56:00.550643 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 23:56:00 crc kubenswrapper[4871]: I1007 23:56:00.617378 4871 scope.go:117] "RemoveContainer" containerID="d4d879ff4ca478c33d8fc835f8d12c1379f8c8ffc4d6bee5aeb7a0dec0959c6c" Oct 07 23:56:00 crc kubenswrapper[4871]: I1007 23:56:00.673149 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f8984b94c-hdzvp"] Oct 07 23:56:00 crc kubenswrapper[4871]: I1007 23:56:00.788628 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 07 23:56:00 crc kubenswrapper[4871]: W1007 23:56:00.805853 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95791187_6a85_4955_997d_920afa9d8adf.slice/crio-57d8b5c7e17de355516162b0af541681dd1edc4e25ddf5d63abf68ad5267268c WatchSource:0}: Error finding container 57d8b5c7e17de355516162b0af541681dd1edc4e25ddf5d63abf68ad5267268c: Status 404 returned error can't find the container with id 57d8b5c7e17de355516162b0af541681dd1edc4e25ddf5d63abf68ad5267268c Oct 07 23:56:01 crc kubenswrapper[4871]: I1007 23:56:01.158081 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"95791187-6a85-4955-997d-920afa9d8adf","Type":"ContainerStarted","Data":"57d8b5c7e17de355516162b0af541681dd1edc4e25ddf5d63abf68ad5267268c"} Oct 07 23:56:01 crc kubenswrapper[4871]: I1007 23:56:01.175969 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e21ff4e8-d971-4e36-bdb5-ea2e5364326e","Type":"ContainerStarted","Data":"1b92916a775f27931d2940397045a402326fecbb188881f0a0d6f70ac68bf9cc"} Oct 07 23:56:01 crc kubenswrapper[4871]: I1007 23:56:01.202038 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"44e3376b-cce2-4e65-a494-c17253d07c09","Type":"ContainerStarted","Data":"6284dbcda2452d50c3b687f8329e9e19c08741f585c6a7668525b3db512cce08"} Oct 07 23:56:01 crc kubenswrapper[4871]: I1007 23:56:01.237004 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" event={"ID":"65a57c8a-e487-4924-bcfe-8fe57300dc57","Type":"ContainerStarted","Data":"77374338168ec47fe9a4e05ff85e8878aa79395e2108e7ec5f5391cd15891574"} Oct 07 23:56:01 crc kubenswrapper[4871]: E1007 23:56:01.493457 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65a57c8a_e487_4924_bcfe_8fe57300dc57.slice/crio-a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65a57c8a_e487_4924_bcfe_8fe57300dc57.slice/crio-conmon-a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3.scope\": RecentStats: unable to find data in memory cache]" Oct 07 23:56:01 crc kubenswrapper[4871]: I1007 23:56:01.982027 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:56:01 crc kubenswrapper[4871]: E1007 23:56:01.982775 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:56:02 crc kubenswrapper[4871]: I1007 23:56:02.259683 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e21ff4e8-d971-4e36-bdb5-ea2e5364326e","Type":"ContainerStarted","Data":"51cd9119770a195441913aaee30bb066672de2157d0b58e97e3c2e223979848b"} Oct 07 23:56:02 crc kubenswrapper[4871]: I1007 23:56:02.262766 4871 generic.go:334] "Generic (PLEG): container finished" podID="65a57c8a-e487-4924-bcfe-8fe57300dc57" containerID="a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3" exitCode=0 Oct 07 23:56:02 crc kubenswrapper[4871]: I1007 23:56:02.262850 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" event={"ID":"65a57c8a-e487-4924-bcfe-8fe57300dc57","Type":"ContainerDied","Data":"a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3"} Oct 07 23:56:02 crc kubenswrapper[4871]: I1007 23:56:02.267866 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"95791187-6a85-4955-997d-920afa9d8adf","Type":"ContainerStarted","Data":"57f60a7b9323cbfa38757a867b1321174b8d8b752d161c09422c179644e802fb"} Oct 07 23:56:02 crc kubenswrapper[4871]: I1007 23:56:02.267929 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"95791187-6a85-4955-997d-920afa9d8adf","Type":"ContainerStarted","Data":"e54823cf083ce2d06fdd76c57fc57eb963c2bc95bce40e1e943cbec510fb0c37"} Oct 07 23:56:02 crc kubenswrapper[4871]: I1007 23:56:02.268015 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 07 23:56:02 crc kubenswrapper[4871]: I1007 23:56:02.312964 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.312947169 podStartE2EDuration="3.312947169s" podCreationTimestamp="2025-10-07 23:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:56:02.3057144 +0000 UTC m=+6436.108412473" watchObservedRunningTime="2025-10-07 23:56:02.312947169 +0000 UTC m=+6436.115645232" Oct 07 23:56:03 crc kubenswrapper[4871]: I1007 23:56:03.280380 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e21ff4e8-d971-4e36-bdb5-ea2e5364326e","Type":"ContainerStarted","Data":"aafaa38741d546a8a697aad904149e4462ce3537e6add3c3794191378a31702c"} Oct 07 23:56:03 crc kubenswrapper[4871]: I1007 23:56:03.291955 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" event={"ID":"65a57c8a-e487-4924-bcfe-8fe57300dc57","Type":"ContainerStarted","Data":"2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6"} Oct 07 23:56:03 crc kubenswrapper[4871]: I1007 23:56:03.292292 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:56:03 crc kubenswrapper[4871]: I1007 23:56:03.303309 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.425755147 podStartE2EDuration="4.303294345s" podCreationTimestamp="2025-10-07 23:55:59 +0000 UTC" firstStartedPulling="2025-10-07 23:56:00.437531428 +0000 UTC m=+6434.240229491" lastFinishedPulling="2025-10-07 23:56:01.315070616 +0000 UTC m=+6435.117768689" observedRunningTime="2025-10-07 23:56:03.298378286 +0000 UTC m=+6437.101076369" watchObservedRunningTime="2025-10-07 23:56:03.303294345 +0000 UTC m=+6437.105992418" Oct 07 23:56:03 crc kubenswrapper[4871]: I1007 23:56:03.332173 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" podStartSLOduration=4.332154962 podStartE2EDuration="4.332154962s" podCreationTimestamp="2025-10-07 23:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:56:03.323454064 +0000 UTC m=+6437.126152137" watchObservedRunningTime="2025-10-07 23:56:03.332154962 +0000 UTC m=+6437.134853035" Oct 07 23:56:08 crc kubenswrapper[4871]: I1007 23:56:08.364863 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"44e3376b-cce2-4e65-a494-c17253d07c09","Type":"ContainerStarted","Data":"b630e29dad96f07764bda5e0aadf5baa36aa2f15c8aed1dbd482c8c96fa70476"} Oct 07 23:56:08 crc kubenswrapper[4871]: I1007 23:56:08.365509 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"44e3376b-cce2-4e65-a494-c17253d07c09","Type":"ContainerStarted","Data":"940ee6e79ea37941446625afaa64496fa7da2207d2f498bf529d71f1b7de2631"} Oct 07 23:56:08 crc kubenswrapper[4871]: I1007 23:56:08.403104 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.722747609 podStartE2EDuration="9.403077688s" podCreationTimestamp="2025-10-07 23:55:59 +0000 UTC" firstStartedPulling="2025-10-07 23:56:00.579677717 +0000 UTC m=+6434.382375800" lastFinishedPulling="2025-10-07 23:56:07.260007796 +0000 UTC m=+6441.062705879" observedRunningTime="2025-10-07 23:56:08.39170372 +0000 UTC m=+6442.194401833" watchObservedRunningTime="2025-10-07 23:56:08.403077688 +0000 UTC m=+6442.205775771" Oct 07 23:56:09 crc kubenswrapper[4871]: I1007 23:56:09.802314 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 07 23:56:09 crc kubenswrapper[4871]: I1007 23:56:09.806667 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 07 23:56:09 crc kubenswrapper[4871]: I1007 23:56:09.876464 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:56:09 crc kubenswrapper[4871]: I1007 23:56:09.982222 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd78b65c9-6mlmx"] Oct 07 23:56:09 crc kubenswrapper[4871]: I1007 23:56:09.982583 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" podUID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" containerName="dnsmasq-dns" containerID="cri-o://df1d813d2ce9d4336674ddcd51960dad2c4c71b3a17fe0f0baef521b98e1edbd" gracePeriod=10 Oct 07 23:56:10 crc kubenswrapper[4871]: I1007 23:56:10.331075 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" podUID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.76:5353: connect: connection refused" Oct 07 23:56:10 crc kubenswrapper[4871]: I1007 23:56:10.384328 4871 generic.go:334] "Generic (PLEG): container finished" podID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" containerID="df1d813d2ce9d4336674ddcd51960dad2c4c71b3a17fe0f0baef521b98e1edbd" exitCode=0 Oct 07 23:56:10 crc kubenswrapper[4871]: I1007 23:56:10.384574 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" event={"ID":"15b9cadf-0dd4-4f1d-a372-84dee39459cd","Type":"ContainerDied","Data":"df1d813d2ce9d4336674ddcd51960dad2c4c71b3a17fe0f0baef521b98e1edbd"} Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.121150 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.224571 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-nb\") pod \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.224650 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6spq\" (UniqueName: \"kubernetes.io/projected/15b9cadf-0dd4-4f1d-a372-84dee39459cd-kube-api-access-z6spq\") pod \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.224746 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-dns-svc\") pod \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.224816 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-sb\") pod \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.224919 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-config\") pod \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\" (UID: \"15b9cadf-0dd4-4f1d-a372-84dee39459cd\") " Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.234046 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15b9cadf-0dd4-4f1d-a372-84dee39459cd-kube-api-access-z6spq" (OuterVolumeSpecName: "kube-api-access-z6spq") pod "15b9cadf-0dd4-4f1d-a372-84dee39459cd" (UID: "15b9cadf-0dd4-4f1d-a372-84dee39459cd"). InnerVolumeSpecName "kube-api-access-z6spq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.303150 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "15b9cadf-0dd4-4f1d-a372-84dee39459cd" (UID: "15b9cadf-0dd4-4f1d-a372-84dee39459cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.314912 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-config" (OuterVolumeSpecName: "config") pod "15b9cadf-0dd4-4f1d-a372-84dee39459cd" (UID: "15b9cadf-0dd4-4f1d-a372-84dee39459cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.315286 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "15b9cadf-0dd4-4f1d-a372-84dee39459cd" (UID: "15b9cadf-0dd4-4f1d-a372-84dee39459cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.327630 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.327664 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.327680 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6spq\" (UniqueName: \"kubernetes.io/projected/15b9cadf-0dd4-4f1d-a372-84dee39459cd-kube-api-access-z6spq\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.327691 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.332591 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "15b9cadf-0dd4-4f1d-a372-84dee39459cd" (UID: "15b9cadf-0dd4-4f1d-a372-84dee39459cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.394045 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" event={"ID":"15b9cadf-0dd4-4f1d-a372-84dee39459cd","Type":"ContainerDied","Data":"17997db74971439eb59c77b73d35454f8a2153e3ad89b0d999319ab9976b62a4"} Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.394111 4871 scope.go:117] "RemoveContainer" containerID="df1d813d2ce9d4336674ddcd51960dad2c4c71b3a17fe0f0baef521b98e1edbd" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.394121 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd78b65c9-6mlmx" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.417604 4871 scope.go:117] "RemoveContainer" containerID="74e6dd2b302ee5246afaf7bc49ad6c5f82be1a5b5b12c8bfa5d797a856f9c1c0" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.432083 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd78b65c9-6mlmx"] Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.432574 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15b9cadf-0dd4-4f1d-a372-84dee39459cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:11 crc kubenswrapper[4871]: I1007 23:56:11.442101 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bd78b65c9-6mlmx"] Oct 07 23:56:12 crc kubenswrapper[4871]: I1007 23:56:12.873472 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:56:12 crc kubenswrapper[4871]: I1007 23:56:12.874155 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="ceilometer-central-agent" containerID="cri-o://780c4a6b26a5c4183e9a1858ae12d291a47de42e792060b23c0c4c6101a75c91" gracePeriod=30 Oct 07 23:56:12 crc kubenswrapper[4871]: I1007 23:56:12.874179 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="proxy-httpd" containerID="cri-o://e877c05868f81f1a524cc43a685ca654e4ff009aa5d3785794b1f2b5b908f7ac" gracePeriod=30 Oct 07 23:56:12 crc kubenswrapper[4871]: I1007 23:56:12.874264 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="ceilometer-notification-agent" containerID="cri-o://d4a709874a19f6598ca7556edc2929abf9f7375e8d7fbdd47e85dc3fe232a8b9" gracePeriod=30 Oct 07 23:56:12 crc kubenswrapper[4871]: I1007 23:56:12.874260 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="sg-core" containerID="cri-o://7161a6179779a43091b0473bfee661ee11ee7a917eaace0d1d7008ce37e83fe8" gracePeriod=30 Oct 07 23:56:12 crc kubenswrapper[4871]: I1007 23:56:12.983152 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:56:12 crc kubenswrapper[4871]: E1007 23:56:12.983377 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:56:13 crc kubenswrapper[4871]: I1007 23:56:13.001824 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" path="/var/lib/kubelet/pods/15b9cadf-0dd4-4f1d-a372-84dee39459cd/volumes" Oct 07 23:56:13 crc kubenswrapper[4871]: I1007 23:56:13.415529 4871 generic.go:334] "Generic (PLEG): container finished" podID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerID="e877c05868f81f1a524cc43a685ca654e4ff009aa5d3785794b1f2b5b908f7ac" exitCode=0 Oct 07 23:56:13 crc kubenswrapper[4871]: I1007 23:56:13.415559 4871 generic.go:334] "Generic (PLEG): container finished" podID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerID="7161a6179779a43091b0473bfee661ee11ee7a917eaace0d1d7008ce37e83fe8" exitCode=2 Oct 07 23:56:13 crc kubenswrapper[4871]: I1007 23:56:13.415567 4871 generic.go:334] "Generic (PLEG): container finished" podID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerID="780c4a6b26a5c4183e9a1858ae12d291a47de42e792060b23c0c4c6101a75c91" exitCode=0 Oct 07 23:56:13 crc kubenswrapper[4871]: I1007 23:56:13.415585 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerDied","Data":"e877c05868f81f1a524cc43a685ca654e4ff009aa5d3785794b1f2b5b908f7ac"} Oct 07 23:56:13 crc kubenswrapper[4871]: I1007 23:56:13.415610 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerDied","Data":"7161a6179779a43091b0473bfee661ee11ee7a917eaace0d1d7008ce37e83fe8"} Oct 07 23:56:13 crc kubenswrapper[4871]: I1007 23:56:13.415620 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerDied","Data":"780c4a6b26a5c4183e9a1858ae12d291a47de42e792060b23c0c4c6101a75c91"} Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.473062 4871 generic.go:334] "Generic (PLEG): container finished" podID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerID="d4a709874a19f6598ca7556edc2929abf9f7375e8d7fbdd47e85dc3fe232a8b9" exitCode=0 Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.473110 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerDied","Data":"d4a709874a19f6598ca7556edc2929abf9f7375e8d7fbdd47e85dc3fe232a8b9"} Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.785448 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.875787 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc4vx\" (UniqueName: \"kubernetes.io/projected/2f03be08-f24c-4da2-9238-191cf91ed53e-kube-api-access-wc4vx\") pod \"2f03be08-f24c-4da2-9238-191cf91ed53e\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.875931 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-sg-core-conf-yaml\") pod \"2f03be08-f24c-4da2-9238-191cf91ed53e\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.875968 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-config-data\") pod \"2f03be08-f24c-4da2-9238-191cf91ed53e\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.876030 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-log-httpd\") pod \"2f03be08-f24c-4da2-9238-191cf91ed53e\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.876081 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-run-httpd\") pod \"2f03be08-f24c-4da2-9238-191cf91ed53e\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.876143 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-combined-ca-bundle\") pod \"2f03be08-f24c-4da2-9238-191cf91ed53e\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.876323 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-scripts\") pod \"2f03be08-f24c-4da2-9238-191cf91ed53e\" (UID: \"2f03be08-f24c-4da2-9238-191cf91ed53e\") " Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.878148 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2f03be08-f24c-4da2-9238-191cf91ed53e" (UID: "2f03be08-f24c-4da2-9238-191cf91ed53e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.879115 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2f03be08-f24c-4da2-9238-191cf91ed53e" (UID: "2f03be08-f24c-4da2-9238-191cf91ed53e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.897019 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f03be08-f24c-4da2-9238-191cf91ed53e-kube-api-access-wc4vx" (OuterVolumeSpecName: "kube-api-access-wc4vx") pod "2f03be08-f24c-4da2-9238-191cf91ed53e" (UID: "2f03be08-f24c-4da2-9238-191cf91ed53e"). InnerVolumeSpecName "kube-api-access-wc4vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.913939 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-scripts" (OuterVolumeSpecName: "scripts") pod "2f03be08-f24c-4da2-9238-191cf91ed53e" (UID: "2f03be08-f24c-4da2-9238-191cf91ed53e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.987708 4871 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.987760 4871 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f03be08-f24c-4da2-9238-191cf91ed53e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.987773 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:17 crc kubenswrapper[4871]: I1007 23:56:17.987785 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc4vx\" (UniqueName: \"kubernetes.io/projected/2f03be08-f24c-4da2-9238-191cf91ed53e-kube-api-access-wc4vx\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.032479 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2f03be08-f24c-4da2-9238-191cf91ed53e" (UID: "2f03be08-f24c-4da2-9238-191cf91ed53e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.057461 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f03be08-f24c-4da2-9238-191cf91ed53e" (UID: "2f03be08-f24c-4da2-9238-191cf91ed53e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.089645 4871 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.089679 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.106305 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-config-data" (OuterVolumeSpecName: "config-data") pod "2f03be08-f24c-4da2-9238-191cf91ed53e" (UID: "2f03be08-f24c-4da2-9238-191cf91ed53e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.191656 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f03be08-f24c-4da2-9238-191cf91ed53e-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.484265 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f03be08-f24c-4da2-9238-191cf91ed53e","Type":"ContainerDied","Data":"04c568746fcc5946ccd6da8f62bf7fa4c1f17ac083c60ee29ec2396915cab846"} Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.484314 4871 scope.go:117] "RemoveContainer" containerID="e877c05868f81f1a524cc43a685ca654e4ff009aa5d3785794b1f2b5b908f7ac" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.484429 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.530116 4871 scope.go:117] "RemoveContainer" containerID="7161a6179779a43091b0473bfee661ee11ee7a917eaace0d1d7008ce37e83fe8" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.531357 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.551967 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.566807 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:56:18 crc kubenswrapper[4871]: E1007 23:56:18.567338 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" containerName="dnsmasq-dns" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.567358 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" containerName="dnsmasq-dns" Oct 07 23:56:18 crc kubenswrapper[4871]: E1007 23:56:18.567368 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="proxy-httpd" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.567375 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="proxy-httpd" Oct 07 23:56:18 crc kubenswrapper[4871]: E1007 23:56:18.567403 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="sg-core" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.567411 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="sg-core" Oct 07 23:56:18 crc kubenswrapper[4871]: E1007 23:56:18.567442 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="ceilometer-central-agent" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.567449 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="ceilometer-central-agent" Oct 07 23:56:18 crc kubenswrapper[4871]: E1007 23:56:18.567462 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" containerName="init" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.567468 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" containerName="init" Oct 07 23:56:18 crc kubenswrapper[4871]: E1007 23:56:18.567481 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="ceilometer-notification-agent" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.567488 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="ceilometer-notification-agent" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.567910 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="ceilometer-central-agent" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.567973 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="proxy-httpd" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.568002 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="sg-core" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.568036 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" containerName="ceilometer-notification-agent" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.568060 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="15b9cadf-0dd4-4f1d-a372-84dee39459cd" containerName="dnsmasq-dns" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.575121 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.579605 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.579891 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.584415 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.602142 4871 scope.go:117] "RemoveContainer" containerID="d4a709874a19f6598ca7556edc2929abf9f7375e8d7fbdd47e85dc3fe232a8b9" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.628105 4871 scope.go:117] "RemoveContainer" containerID="780c4a6b26a5c4183e9a1858ae12d291a47de42e792060b23c0c4c6101a75c91" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.702418 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/304c982b-0677-407f-88c4-4efbbc47a363-log-httpd\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.702524 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-scripts\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.702564 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/304c982b-0677-407f-88c4-4efbbc47a363-run-httpd\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.702586 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms8h7\" (UniqueName: \"kubernetes.io/projected/304c982b-0677-407f-88c4-4efbbc47a363-kube-api-access-ms8h7\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.702615 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-config-data\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.702643 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.702664 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.804771 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/304c982b-0677-407f-88c4-4efbbc47a363-log-httpd\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.804906 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-scripts\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.804945 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/304c982b-0677-407f-88c4-4efbbc47a363-run-httpd\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.804964 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms8h7\" (UniqueName: \"kubernetes.io/projected/304c982b-0677-407f-88c4-4efbbc47a363-kube-api-access-ms8h7\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.804999 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-config-data\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.805030 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.805051 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.805964 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/304c982b-0677-407f-88c4-4efbbc47a363-run-httpd\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.806210 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/304c982b-0677-407f-88c4-4efbbc47a363-log-httpd\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.816435 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.816591 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.816727 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-config-data\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.816877 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/304c982b-0677-407f-88c4-4efbbc47a363-scripts\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.821437 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms8h7\" (UniqueName: \"kubernetes.io/projected/304c982b-0677-407f-88c4-4efbbc47a363-kube-api-access-ms8h7\") pod \"ceilometer-0\" (UID: \"304c982b-0677-407f-88c4-4efbbc47a363\") " pod="openstack/ceilometer-0" Oct 07 23:56:18 crc kubenswrapper[4871]: I1007 23:56:18.899659 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 23:56:19 crc kubenswrapper[4871]: I1007 23:56:19.024919 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f03be08-f24c-4da2-9238-191cf91ed53e" path="/var/lib/kubelet/pods/2f03be08-f24c-4da2-9238-191cf91ed53e/volumes" Oct 07 23:56:19 crc kubenswrapper[4871]: I1007 23:56:19.419219 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 23:56:19 crc kubenswrapper[4871]: I1007 23:56:19.494056 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"304c982b-0677-407f-88c4-4efbbc47a363","Type":"ContainerStarted","Data":"b7bf4052fa5d9353366b30ec6a256fe370140a9cc4029f0a4c96b7669ca66703"} Oct 07 23:56:20 crc kubenswrapper[4871]: I1007 23:56:20.506479 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"304c982b-0677-407f-88c4-4efbbc47a363","Type":"ContainerStarted","Data":"246d7dc73c97d0c52af91394e44ca655d2870af8acc5ec753bd9a311cbd5546d"} Oct 07 23:56:21 crc kubenswrapper[4871]: I1007 23:56:21.354513 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 07 23:56:21 crc kubenswrapper[4871]: I1007 23:56:21.462340 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 07 23:56:21 crc kubenswrapper[4871]: I1007 23:56:21.518383 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"304c982b-0677-407f-88c4-4efbbc47a363","Type":"ContainerStarted","Data":"eaf9248839051700b6b4451a0c66506b47d4af0a184eaf5fcb4bfafe6a638652"} Oct 07 23:56:21 crc kubenswrapper[4871]: I1007 23:56:21.592478 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 07 23:56:22 crc kubenswrapper[4871]: I1007 23:56:22.528890 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"304c982b-0677-407f-88c4-4efbbc47a363","Type":"ContainerStarted","Data":"fc7cd49604e25d5d942437ebf794cf49a78f4d9484756daf06fe0fe49e1ea45b"} Oct 07 23:56:23 crc kubenswrapper[4871]: I1007 23:56:23.982612 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:56:23 crc kubenswrapper[4871]: E1007 23:56:23.983628 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:56:24 crc kubenswrapper[4871]: I1007 23:56:24.556965 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"304c982b-0677-407f-88c4-4efbbc47a363","Type":"ContainerStarted","Data":"9abfa693e0970d3172d75dc8ba50b6ef99f4dc456a2825ec3b34123e6db1e523"} Oct 07 23:56:24 crc kubenswrapper[4871]: I1007 23:56:24.557118 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 23:56:24 crc kubenswrapper[4871]: I1007 23:56:24.586249 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.561800401 podStartE2EDuration="6.586227619s" podCreationTimestamp="2025-10-07 23:56:18 +0000 UTC" firstStartedPulling="2025-10-07 23:56:19.426836553 +0000 UTC m=+6453.229534626" lastFinishedPulling="2025-10-07 23:56:23.451263751 +0000 UTC m=+6457.253961844" observedRunningTime="2025-10-07 23:56:24.584961556 +0000 UTC m=+6458.387659669" watchObservedRunningTime="2025-10-07 23:56:24.586227619 +0000 UTC m=+6458.388925712" Oct 07 23:56:37 crc kubenswrapper[4871]: I1007 23:56:37.983192 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:56:37 crc kubenswrapper[4871]: E1007 23:56:37.984003 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:56:48 crc kubenswrapper[4871]: I1007 23:56:48.908203 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 23:56:50 crc kubenswrapper[4871]: I1007 23:56:50.982568 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:56:50 crc kubenswrapper[4871]: E1007 23:56:50.983362 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:57:01 crc kubenswrapper[4871]: I1007 23:57:01.982746 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:57:01 crc kubenswrapper[4871]: E1007 23:57:01.983881 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:57:08 crc kubenswrapper[4871]: I1007 23:57:08.960624 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6db86745-n29gb"] Oct 07 23:57:08 crc kubenswrapper[4871]: I1007 23:57:08.966223 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:08 crc kubenswrapper[4871]: I1007 23:57:08.969163 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.012547 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6db86745-n29gb"] Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.068681 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-dns-svc\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.068899 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-config\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.069121 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-openstack-cell1\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.069387 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw9ml\" (UniqueName: \"kubernetes.io/projected/d6430e30-d77c-4995-8d01-4213008b6d75-kube-api-access-rw9ml\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.069460 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-sb\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.069662 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-nb\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.172206 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw9ml\" (UniqueName: \"kubernetes.io/projected/d6430e30-d77c-4995-8d01-4213008b6d75-kube-api-access-rw9ml\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.172273 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-sb\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.172351 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-nb\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.173424 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-sb\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.173486 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-nb\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.173639 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-dns-svc\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.175749 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-dns-svc\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.175761 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-config\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.176034 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-openstack-cell1\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.176560 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-config\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.176733 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-openstack-cell1\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.194007 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw9ml\" (UniqueName: \"kubernetes.io/projected/d6430e30-d77c-4995-8d01-4213008b6d75-kube-api-access-rw9ml\") pod \"dnsmasq-dns-7c6db86745-n29gb\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.290726 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:09 crc kubenswrapper[4871]: I1007 23:57:09.768003 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6db86745-n29gb"] Oct 07 23:57:10 crc kubenswrapper[4871]: I1007 23:57:10.096579 4871 generic.go:334] "Generic (PLEG): container finished" podID="d6430e30-d77c-4995-8d01-4213008b6d75" containerID="c8e8953761ae375ba43d3df8b077c673540797f48a641b6fcfba1fc17229a5d2" exitCode=0 Oct 07 23:57:10 crc kubenswrapper[4871]: I1007 23:57:10.097010 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" event={"ID":"d6430e30-d77c-4995-8d01-4213008b6d75","Type":"ContainerDied","Data":"c8e8953761ae375ba43d3df8b077c673540797f48a641b6fcfba1fc17229a5d2"} Oct 07 23:57:10 crc kubenswrapper[4871]: I1007 23:57:10.097039 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" event={"ID":"d6430e30-d77c-4995-8d01-4213008b6d75","Type":"ContainerStarted","Data":"3064168ed1acd5166401ccd12d3c580e02fbc97ef2c4b83c9f258d7662850cac"} Oct 07 23:57:11 crc kubenswrapper[4871]: I1007 23:57:11.108467 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" event={"ID":"d6430e30-d77c-4995-8d01-4213008b6d75","Type":"ContainerStarted","Data":"05138fb7991e3925426221a280c0ee8e16865902438785966cd6a5b83dd9c689"} Oct 07 23:57:11 crc kubenswrapper[4871]: I1007 23:57:11.109189 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:11 crc kubenswrapper[4871]: I1007 23:57:11.137192 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" podStartSLOduration=3.137176433 podStartE2EDuration="3.137176433s" podCreationTimestamp="2025-10-07 23:57:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:57:11.13055149 +0000 UTC m=+6504.933249563" watchObservedRunningTime="2025-10-07 23:57:11.137176433 +0000 UTC m=+6504.939874496" Oct 07 23:57:12 crc kubenswrapper[4871]: I1007 23:57:12.983587 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:57:12 crc kubenswrapper[4871]: E1007 23:57:12.984576 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.293085 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.379721 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f8984b94c-hdzvp"] Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.380134 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" podUID="65a57c8a-e487-4924-bcfe-8fe57300dc57" containerName="dnsmasq-dns" containerID="cri-o://2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6" gracePeriod=10 Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.672532 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-855fbfb687-7pd4r"] Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.674346 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.694406 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-855fbfb687-7pd4r"] Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.828215 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-dns-svc\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.828514 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz92h\" (UniqueName: \"kubernetes.io/projected/ad432e67-ef64-408a-9b6b-4b76a60d308e-kube-api-access-kz92h\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.828549 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-ovsdbserver-nb\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.828595 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-ovsdbserver-sb\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.828691 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-openstack-cell1\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.828731 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-config\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.931394 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-openstack-cell1\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.931480 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-config\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.931618 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-dns-svc\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.931651 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz92h\" (UniqueName: \"kubernetes.io/projected/ad432e67-ef64-408a-9b6b-4b76a60d308e-kube-api-access-kz92h\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.931673 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-ovsdbserver-nb\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.931722 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-ovsdbserver-sb\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.932653 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-openstack-cell1\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.932929 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-ovsdbserver-sb\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.933553 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-dns-svc\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.934410 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-config\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.934535 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad432e67-ef64-408a-9b6b-4b76a60d308e-ovsdbserver-nb\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.959253 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz92h\" (UniqueName: \"kubernetes.io/projected/ad432e67-ef64-408a-9b6b-4b76a60d308e-kube-api-access-kz92h\") pod \"dnsmasq-dns-855fbfb687-7pd4r\" (UID: \"ad432e67-ef64-408a-9b6b-4b76a60d308e\") " pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:19 crc kubenswrapper[4871]: I1007 23:57:19.994830 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.114026 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.224152 4871 generic.go:334] "Generic (PLEG): container finished" podID="65a57c8a-e487-4924-bcfe-8fe57300dc57" containerID="2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6" exitCode=0 Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.224337 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" event={"ID":"65a57c8a-e487-4924-bcfe-8fe57300dc57","Type":"ContainerDied","Data":"2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6"} Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.224442 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.224464 4871 scope.go:117] "RemoveContainer" containerID="2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.224445 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" event={"ID":"65a57c8a-e487-4924-bcfe-8fe57300dc57","Type":"ContainerDied","Data":"77374338168ec47fe9a4e05ff85e8878aa79395e2108e7ec5f5391cd15891574"} Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.248596 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxrdl\" (UniqueName: \"kubernetes.io/projected/65a57c8a-e487-4924-bcfe-8fe57300dc57-kube-api-access-bxrdl\") pod \"65a57c8a-e487-4924-bcfe-8fe57300dc57\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.248920 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-dns-svc\") pod \"65a57c8a-e487-4924-bcfe-8fe57300dc57\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.248976 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-sb\") pod \"65a57c8a-e487-4924-bcfe-8fe57300dc57\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.249871 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-config\") pod \"65a57c8a-e487-4924-bcfe-8fe57300dc57\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.249988 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-nb\") pod \"65a57c8a-e487-4924-bcfe-8fe57300dc57\" (UID: \"65a57c8a-e487-4924-bcfe-8fe57300dc57\") " Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.253106 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65a57c8a-e487-4924-bcfe-8fe57300dc57-kube-api-access-bxrdl" (OuterVolumeSpecName: "kube-api-access-bxrdl") pod "65a57c8a-e487-4924-bcfe-8fe57300dc57" (UID: "65a57c8a-e487-4924-bcfe-8fe57300dc57"). InnerVolumeSpecName "kube-api-access-bxrdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.272319 4871 scope.go:117] "RemoveContainer" containerID="a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.304128 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-config" (OuterVolumeSpecName: "config") pod "65a57c8a-e487-4924-bcfe-8fe57300dc57" (UID: "65a57c8a-e487-4924-bcfe-8fe57300dc57"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.321225 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "65a57c8a-e487-4924-bcfe-8fe57300dc57" (UID: "65a57c8a-e487-4924-bcfe-8fe57300dc57"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.343385 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "65a57c8a-e487-4924-bcfe-8fe57300dc57" (UID: "65a57c8a-e487-4924-bcfe-8fe57300dc57"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.348452 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "65a57c8a-e487-4924-bcfe-8fe57300dc57" (UID: "65a57c8a-e487-4924-bcfe-8fe57300dc57"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.357342 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.357545 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.357558 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.357565 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65a57c8a-e487-4924-bcfe-8fe57300dc57-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.357575 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxrdl\" (UniqueName: \"kubernetes.io/projected/65a57c8a-e487-4924-bcfe-8fe57300dc57-kube-api-access-bxrdl\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.422947 4871 scope.go:117] "RemoveContainer" containerID="2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6" Oct 07 23:57:20 crc kubenswrapper[4871]: E1007 23:57:20.423350 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6\": container with ID starting with 2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6 not found: ID does not exist" containerID="2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.423374 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6"} err="failed to get container status \"2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6\": rpc error: code = NotFound desc = could not find container \"2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6\": container with ID starting with 2a16666c5812413f3fdc868ed025fa7061c69f9b5fb050797953afbbbfae6bd6 not found: ID does not exist" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.423393 4871 scope.go:117] "RemoveContainer" containerID="a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3" Oct 07 23:57:20 crc kubenswrapper[4871]: E1007 23:57:20.423656 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3\": container with ID starting with a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3 not found: ID does not exist" containerID="a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.423702 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3"} err="failed to get container status \"a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3\": rpc error: code = NotFound desc = could not find container \"a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3\": container with ID starting with a982ea96cd481d8732b5fae87580ba4becdcf946dd1c3e9ac5e5bb0b75cc77d3 not found: ID does not exist" Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.575129 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f8984b94c-hdzvp"] Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.593188 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f8984b94c-hdzvp"] Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.621836 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-855fbfb687-7pd4r"] Oct 07 23:57:20 crc kubenswrapper[4871]: I1007 23:57:20.997647 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65a57c8a-e487-4924-bcfe-8fe57300dc57" path="/var/lib/kubelet/pods/65a57c8a-e487-4924-bcfe-8fe57300dc57/volumes" Oct 07 23:57:21 crc kubenswrapper[4871]: I1007 23:57:21.234937 4871 generic.go:334] "Generic (PLEG): container finished" podID="ad432e67-ef64-408a-9b6b-4b76a60d308e" containerID="7e39e6b9526733296ddb5e2ae5d3a1c2e2d21838f595000cecc7a9a83b3b5704" exitCode=0 Oct 07 23:57:21 crc kubenswrapper[4871]: I1007 23:57:21.234974 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" event={"ID":"ad432e67-ef64-408a-9b6b-4b76a60d308e","Type":"ContainerDied","Data":"7e39e6b9526733296ddb5e2ae5d3a1c2e2d21838f595000cecc7a9a83b3b5704"} Oct 07 23:57:21 crc kubenswrapper[4871]: I1007 23:57:21.234995 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" event={"ID":"ad432e67-ef64-408a-9b6b-4b76a60d308e","Type":"ContainerStarted","Data":"2c7dd8fcf9aee8528c2280970d65483cf52a085aecdfc3a04e1b57cbf38b27ca"} Oct 07 23:57:22 crc kubenswrapper[4871]: I1007 23:57:22.247259 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" event={"ID":"ad432e67-ef64-408a-9b6b-4b76a60d308e","Type":"ContainerStarted","Data":"0ee2e7c7769bb5245b21ee1cb46db69dd006483f87dff0bb7a07de11839da60d"} Oct 07 23:57:22 crc kubenswrapper[4871]: I1007 23:57:22.248642 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:22 crc kubenswrapper[4871]: I1007 23:57:22.268713 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" podStartSLOduration=3.268695254 podStartE2EDuration="3.268695254s" podCreationTimestamp="2025-10-07 23:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 23:57:22.264247818 +0000 UTC m=+6516.066945911" watchObservedRunningTime="2025-10-07 23:57:22.268695254 +0000 UTC m=+6516.071393327" Oct 07 23:57:24 crc kubenswrapper[4871]: I1007 23:57:24.874949 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-f8984b94c-hdzvp" podUID="65a57c8a-e487-4924-bcfe-8fe57300dc57" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.141:5353: i/o timeout" Oct 07 23:57:26 crc kubenswrapper[4871]: I1007 23:57:26.991650 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:57:26 crc kubenswrapper[4871]: E1007 23:57:26.992550 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:57:29 crc kubenswrapper[4871]: I1007 23:57:29.997000 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-855fbfb687-7pd4r" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.071177 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6db86745-n29gb"] Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.071467 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" podUID="d6430e30-d77c-4995-8d01-4213008b6d75" containerName="dnsmasq-dns" containerID="cri-o://05138fb7991e3925426221a280c0ee8e16865902438785966cd6a5b83dd9c689" gracePeriod=10 Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.359358 4871 generic.go:334] "Generic (PLEG): container finished" podID="d6430e30-d77c-4995-8d01-4213008b6d75" containerID="05138fb7991e3925426221a280c0ee8e16865902438785966cd6a5b83dd9c689" exitCode=0 Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.359408 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" event={"ID":"d6430e30-d77c-4995-8d01-4213008b6d75","Type":"ContainerDied","Data":"05138fb7991e3925426221a280c0ee8e16865902438785966cd6a5b83dd9c689"} Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.671229 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.825920 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-sb\") pod \"d6430e30-d77c-4995-8d01-4213008b6d75\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.825996 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-config\") pod \"d6430e30-d77c-4995-8d01-4213008b6d75\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.826106 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw9ml\" (UniqueName: \"kubernetes.io/projected/d6430e30-d77c-4995-8d01-4213008b6d75-kube-api-access-rw9ml\") pod \"d6430e30-d77c-4995-8d01-4213008b6d75\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.826144 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-nb\") pod \"d6430e30-d77c-4995-8d01-4213008b6d75\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.826892 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-dns-svc\") pod \"d6430e30-d77c-4995-8d01-4213008b6d75\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.826968 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-openstack-cell1\") pod \"d6430e30-d77c-4995-8d01-4213008b6d75\" (UID: \"d6430e30-d77c-4995-8d01-4213008b6d75\") " Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.831384 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6430e30-d77c-4995-8d01-4213008b6d75-kube-api-access-rw9ml" (OuterVolumeSpecName: "kube-api-access-rw9ml") pod "d6430e30-d77c-4995-8d01-4213008b6d75" (UID: "d6430e30-d77c-4995-8d01-4213008b6d75"). InnerVolumeSpecName "kube-api-access-rw9ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.879094 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d6430e30-d77c-4995-8d01-4213008b6d75" (UID: "d6430e30-d77c-4995-8d01-4213008b6d75"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.888406 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-config" (OuterVolumeSpecName: "config") pod "d6430e30-d77c-4995-8d01-4213008b6d75" (UID: "d6430e30-d77c-4995-8d01-4213008b6d75"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.893230 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d6430e30-d77c-4995-8d01-4213008b6d75" (UID: "d6430e30-d77c-4995-8d01-4213008b6d75"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.900035 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d6430e30-d77c-4995-8d01-4213008b6d75" (UID: "d6430e30-d77c-4995-8d01-4213008b6d75"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.920543 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "d6430e30-d77c-4995-8d01-4213008b6d75" (UID: "d6430e30-d77c-4995-8d01-4213008b6d75"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.930309 4871 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-config\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.930348 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw9ml\" (UniqueName: \"kubernetes.io/projected/d6430e30-d77c-4995-8d01-4213008b6d75-kube-api-access-rw9ml\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.930359 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.930370 4871 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.930379 4871 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:30 crc kubenswrapper[4871]: I1007 23:57:30.930387 4871 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6430e30-d77c-4995-8d01-4213008b6d75-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 23:57:31 crc kubenswrapper[4871]: I1007 23:57:31.376654 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" event={"ID":"d6430e30-d77c-4995-8d01-4213008b6d75","Type":"ContainerDied","Data":"3064168ed1acd5166401ccd12d3c580e02fbc97ef2c4b83c9f258d7662850cac"} Oct 07 23:57:31 crc kubenswrapper[4871]: I1007 23:57:31.377032 4871 scope.go:117] "RemoveContainer" containerID="05138fb7991e3925426221a280c0ee8e16865902438785966cd6a5b83dd9c689" Oct 07 23:57:31 crc kubenswrapper[4871]: I1007 23:57:31.376700 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6db86745-n29gb" Oct 07 23:57:31 crc kubenswrapper[4871]: I1007 23:57:31.405562 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6db86745-n29gb"] Oct 07 23:57:31 crc kubenswrapper[4871]: I1007 23:57:31.414271 4871 scope.go:117] "RemoveContainer" containerID="c8e8953761ae375ba43d3df8b077c673540797f48a641b6fcfba1fc17229a5d2" Oct 07 23:57:31 crc kubenswrapper[4871]: I1007 23:57:31.415846 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6db86745-n29gb"] Oct 07 23:57:33 crc kubenswrapper[4871]: I1007 23:57:33.010687 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6430e30-d77c-4995-8d01-4213008b6d75" path="/var/lib/kubelet/pods/d6430e30-d77c-4995-8d01-4213008b6d75/volumes" Oct 07 23:57:37 crc kubenswrapper[4871]: I1007 23:57:37.982775 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:57:37 crc kubenswrapper[4871]: E1007 23:57:37.984020 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.891312 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr"] Oct 07 23:57:40 crc kubenswrapper[4871]: E1007 23:57:40.893098 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65a57c8a-e487-4924-bcfe-8fe57300dc57" containerName="dnsmasq-dns" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.893138 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="65a57c8a-e487-4924-bcfe-8fe57300dc57" containerName="dnsmasq-dns" Oct 07 23:57:40 crc kubenswrapper[4871]: E1007 23:57:40.893187 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6430e30-d77c-4995-8d01-4213008b6d75" containerName="init" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.893205 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6430e30-d77c-4995-8d01-4213008b6d75" containerName="init" Oct 07 23:57:40 crc kubenswrapper[4871]: E1007 23:57:40.893249 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65a57c8a-e487-4924-bcfe-8fe57300dc57" containerName="init" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.893268 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="65a57c8a-e487-4924-bcfe-8fe57300dc57" containerName="init" Oct 07 23:57:40 crc kubenswrapper[4871]: E1007 23:57:40.893307 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6430e30-d77c-4995-8d01-4213008b6d75" containerName="dnsmasq-dns" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.893326 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6430e30-d77c-4995-8d01-4213008b6d75" containerName="dnsmasq-dns" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.893867 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="65a57c8a-e487-4924-bcfe-8fe57300dc57" containerName="dnsmasq-dns" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.893919 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6430e30-d77c-4995-8d01-4213008b6d75" containerName="dnsmasq-dns" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.897138 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.901042 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.901230 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.901457 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.901646 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.907079 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr"] Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.984425 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.984512 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.984560 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.984649 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktn55\" (UniqueName: \"kubernetes.io/projected/be4cf803-f602-46b3-90ba-e17ed62e248c-kube-api-access-ktn55\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:40 crc kubenswrapper[4871]: I1007 23:57:40.984691 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.086615 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktn55\" (UniqueName: \"kubernetes.io/projected/be4cf803-f602-46b3-90ba-e17ed62e248c-kube-api-access-ktn55\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.086739 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.086842 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.086964 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.087075 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.096752 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.097434 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.102744 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.103730 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.114874 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktn55\" (UniqueName: \"kubernetes.io/projected/be4cf803-f602-46b3-90ba-e17ed62e248c-kube-api-access-ktn55\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.231831 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:57:41 crc kubenswrapper[4871]: I1007 23:57:41.917429 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr"] Oct 07 23:57:41 crc kubenswrapper[4871]: W1007 23:57:41.928040 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe4cf803_f602_46b3_90ba_e17ed62e248c.slice/crio-b8e2133f63a7dce28cddd3d8ad25906d47360b22a84ec5c465b41991db91f0e7 WatchSource:0}: Error finding container b8e2133f63a7dce28cddd3d8ad25906d47360b22a84ec5c465b41991db91f0e7: Status 404 returned error can't find the container with id b8e2133f63a7dce28cddd3d8ad25906d47360b22a84ec5c465b41991db91f0e7 Oct 07 23:57:42 crc kubenswrapper[4871]: I1007 23:57:42.503268 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" event={"ID":"be4cf803-f602-46b3-90ba-e17ed62e248c","Type":"ContainerStarted","Data":"b8e2133f63a7dce28cddd3d8ad25906d47360b22a84ec5c465b41991db91f0e7"} Oct 07 23:57:50 crc kubenswrapper[4871]: I1007 23:57:50.438279 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 23:57:50 crc kubenswrapper[4871]: I1007 23:57:50.983348 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:57:50 crc kubenswrapper[4871]: E1007 23:57:50.984298 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:57:51 crc kubenswrapper[4871]: I1007 23:57:51.628585 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" event={"ID":"be4cf803-f602-46b3-90ba-e17ed62e248c","Type":"ContainerStarted","Data":"201a25d5f8adee49233fa4fc332a48776809eba02f0bf2c3dd379d4cd0980916"} Oct 07 23:57:51 crc kubenswrapper[4871]: I1007 23:57:51.659607 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" podStartSLOduration=3.158668703 podStartE2EDuration="11.659579034s" podCreationTimestamp="2025-10-07 23:57:40 +0000 UTC" firstStartedPulling="2025-10-07 23:57:41.933505217 +0000 UTC m=+6535.736203300" lastFinishedPulling="2025-10-07 23:57:50.434415518 +0000 UTC m=+6544.237113631" observedRunningTime="2025-10-07 23:57:51.648990376 +0000 UTC m=+6545.451688489" watchObservedRunningTime="2025-10-07 23:57:51.659579034 +0000 UTC m=+6545.462277137" Oct 07 23:57:58 crc kubenswrapper[4871]: I1007 23:57:58.064118 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-gq7n9"] Oct 07 23:57:58 crc kubenswrapper[4871]: I1007 23:57:58.077704 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-gq7n9"] Oct 07 23:57:58 crc kubenswrapper[4871]: I1007 23:57:58.996259 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c9dee7b-aba0-4d5d-8562-0550d4db7584" path="/var/lib/kubelet/pods/3c9dee7b-aba0-4d5d-8562-0550d4db7584/volumes" Oct 07 23:58:00 crc kubenswrapper[4871]: I1007 23:58:00.855951 4871 scope.go:117] "RemoveContainer" containerID="a710dcd16e9ce1303317222eba462dce03c60048634e33123a848640a010055f" Oct 07 23:58:03 crc kubenswrapper[4871]: I1007 23:58:03.984098 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:58:03 crc kubenswrapper[4871]: E1007 23:58:03.985037 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:58:04 crc kubenswrapper[4871]: I1007 23:58:04.806865 4871 generic.go:334] "Generic (PLEG): container finished" podID="be4cf803-f602-46b3-90ba-e17ed62e248c" containerID="201a25d5f8adee49233fa4fc332a48776809eba02f0bf2c3dd379d4cd0980916" exitCode=0 Oct 07 23:58:04 crc kubenswrapper[4871]: I1007 23:58:04.807022 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" event={"ID":"be4cf803-f602-46b3-90ba-e17ed62e248c","Type":"ContainerDied","Data":"201a25d5f8adee49233fa4fc332a48776809eba02f0bf2c3dd379d4cd0980916"} Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.384639 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.413009 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-pre-adoption-validation-combined-ca-bundle\") pod \"be4cf803-f602-46b3-90ba-e17ed62e248c\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.413101 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ssh-key\") pod \"be4cf803-f602-46b3-90ba-e17ed62e248c\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.413299 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktn55\" (UniqueName: \"kubernetes.io/projected/be4cf803-f602-46b3-90ba-e17ed62e248c-kube-api-access-ktn55\") pod \"be4cf803-f602-46b3-90ba-e17ed62e248c\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.413393 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ceph\") pod \"be4cf803-f602-46b3-90ba-e17ed62e248c\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.413778 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-inventory\") pod \"be4cf803-f602-46b3-90ba-e17ed62e248c\" (UID: \"be4cf803-f602-46b3-90ba-e17ed62e248c\") " Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.429100 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ceph" (OuterVolumeSpecName: "ceph") pod "be4cf803-f602-46b3-90ba-e17ed62e248c" (UID: "be4cf803-f602-46b3-90ba-e17ed62e248c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.429469 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be4cf803-f602-46b3-90ba-e17ed62e248c-kube-api-access-ktn55" (OuterVolumeSpecName: "kube-api-access-ktn55") pod "be4cf803-f602-46b3-90ba-e17ed62e248c" (UID: "be4cf803-f602-46b3-90ba-e17ed62e248c"). InnerVolumeSpecName "kube-api-access-ktn55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.429700 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "be4cf803-f602-46b3-90ba-e17ed62e248c" (UID: "be4cf803-f602-46b3-90ba-e17ed62e248c"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.459116 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "be4cf803-f602-46b3-90ba-e17ed62e248c" (UID: "be4cf803-f602-46b3-90ba-e17ed62e248c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.469586 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-inventory" (OuterVolumeSpecName: "inventory") pod "be4cf803-f602-46b3-90ba-e17ed62e248c" (UID: "be4cf803-f602-46b3-90ba-e17ed62e248c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.517326 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.517368 4871 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.517383 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.517401 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktn55\" (UniqueName: \"kubernetes.io/projected/be4cf803-f602-46b3-90ba-e17ed62e248c-kube-api-access-ktn55\") on node \"crc\" DevicePath \"\"" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.517413 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be4cf803-f602-46b3-90ba-e17ed62e248c-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.836331 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" event={"ID":"be4cf803-f602-46b3-90ba-e17ed62e248c","Type":"ContainerDied","Data":"b8e2133f63a7dce28cddd3d8ad25906d47360b22a84ec5c465b41991db91f0e7"} Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.836380 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8e2133f63a7dce28cddd3d8ad25906d47360b22a84ec5c465b41991db91f0e7" Oct 07 23:58:06 crc kubenswrapper[4871]: I1007 23:58:06.836947 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr" Oct 07 23:58:10 crc kubenswrapper[4871]: I1007 23:58:10.069709 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-8bf7-account-create-s55bk"] Oct 07 23:58:10 crc kubenswrapper[4871]: I1007 23:58:10.080737 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-8bf7-account-create-s55bk"] Oct 07 23:58:11 crc kubenswrapper[4871]: I1007 23:58:11.010354 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a00b25ff-46ef-4d87-ad93-81543f7851fa" path="/var/lib/kubelet/pods/a00b25ff-46ef-4d87-ad93-81543f7851fa/volumes" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.850157 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6"] Oct 07 23:58:13 crc kubenswrapper[4871]: E1007 23:58:13.851234 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be4cf803-f602-46b3-90ba-e17ed62e248c" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.851258 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="be4cf803-f602-46b3-90ba-e17ed62e248c" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.851673 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="be4cf803-f602-46b3-90ba-e17ed62e248c" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.852994 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.856964 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.859992 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.860059 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.860207 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.871245 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6"] Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.906147 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.906257 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.906345 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.906498 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:13 crc kubenswrapper[4871]: I1007 23:58:13.906598 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8b58\" (UniqueName: \"kubernetes.io/projected/8a39a36c-8c17-4049-99a0-3ef681c082b3-kube-api-access-b8b58\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.009063 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8b58\" (UniqueName: \"kubernetes.io/projected/8a39a36c-8c17-4049-99a0-3ef681c082b3-kube-api-access-b8b58\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.009770 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.010148 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.010249 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.010414 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.016826 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.017420 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.018205 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.018389 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.024331 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8b58\" (UniqueName: \"kubernetes.io/projected/8a39a36c-8c17-4049-99a0-3ef681c082b3-kube-api-access-b8b58\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.191128 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.774966 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6"] Oct 07 23:58:14 crc kubenswrapper[4871]: W1007 23:58:14.786767 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a39a36c_8c17_4049_99a0_3ef681c082b3.slice/crio-189c803ff3ad571a4ff3600411123249b1820bc1be19c0cf56a6684aa0eefb35 WatchSource:0}: Error finding container 189c803ff3ad571a4ff3600411123249b1820bc1be19c0cf56a6684aa0eefb35: Status 404 returned error can't find the container with id 189c803ff3ad571a4ff3600411123249b1820bc1be19c0cf56a6684aa0eefb35 Oct 07 23:58:14 crc kubenswrapper[4871]: I1007 23:58:14.967864 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" event={"ID":"8a39a36c-8c17-4049-99a0-3ef681c082b3","Type":"ContainerStarted","Data":"189c803ff3ad571a4ff3600411123249b1820bc1be19c0cf56a6684aa0eefb35"} Oct 07 23:58:15 crc kubenswrapper[4871]: I1007 23:58:15.980474 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" event={"ID":"8a39a36c-8c17-4049-99a0-3ef681c082b3","Type":"ContainerStarted","Data":"bfe0d5210a49c7f33d9a443c7a9de84809e4a127626addc0ad4d4885795a1a0c"} Oct 07 23:58:16 crc kubenswrapper[4871]: I1007 23:58:16.012117 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" podStartSLOduration=2.564989864 podStartE2EDuration="3.01209712s" podCreationTimestamp="2025-10-07 23:58:13 +0000 UTC" firstStartedPulling="2025-10-07 23:58:14.790166361 +0000 UTC m=+6568.592864434" lastFinishedPulling="2025-10-07 23:58:15.237273607 +0000 UTC m=+6569.039971690" observedRunningTime="2025-10-07 23:58:16.006476413 +0000 UTC m=+6569.809174486" watchObservedRunningTime="2025-10-07 23:58:16.01209712 +0000 UTC m=+6569.814795193" Oct 07 23:58:17 crc kubenswrapper[4871]: I1007 23:58:17.037065 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-x4fxr"] Oct 07 23:58:17 crc kubenswrapper[4871]: I1007 23:58:17.048875 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-x4fxr"] Oct 07 23:58:17 crc kubenswrapper[4871]: I1007 23:58:17.982682 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:58:17 crc kubenswrapper[4871]: E1007 23:58:17.983178 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:58:19 crc kubenswrapper[4871]: I1007 23:58:19.001670 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="555b619b-6b5e-403c-ab31-4484710488a3" path="/var/lib/kubelet/pods/555b619b-6b5e-403c-ab31-4484710488a3/volumes" Oct 07 23:58:27 crc kubenswrapper[4871]: I1007 23:58:27.881370 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t7mjz"] Oct 07 23:58:27 crc kubenswrapper[4871]: I1007 23:58:27.886420 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:27 crc kubenswrapper[4871]: I1007 23:58:27.918052 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7mjz"] Oct 07 23:58:27 crc kubenswrapper[4871]: I1007 23:58:27.993641 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-utilities\") pod \"community-operators-t7mjz\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:27 crc kubenswrapper[4871]: I1007 23:58:27.993758 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-catalog-content\") pod \"community-operators-t7mjz\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:27 crc kubenswrapper[4871]: I1007 23:58:27.993920 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvdqm\" (UniqueName: \"kubernetes.io/projected/caece94b-ab07-469c-bc19-b0afb73a7134-kube-api-access-wvdqm\") pod \"community-operators-t7mjz\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.035040 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-f4cd-account-create-rdx2p"] Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.046093 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-f4cd-account-create-rdx2p"] Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.096047 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvdqm\" (UniqueName: \"kubernetes.io/projected/caece94b-ab07-469c-bc19-b0afb73a7134-kube-api-access-wvdqm\") pod \"community-operators-t7mjz\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.096175 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-utilities\") pod \"community-operators-t7mjz\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.096293 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-catalog-content\") pod \"community-operators-t7mjz\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.096711 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-catalog-content\") pod \"community-operators-t7mjz\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.097692 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-utilities\") pod \"community-operators-t7mjz\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.121487 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvdqm\" (UniqueName: \"kubernetes.io/projected/caece94b-ab07-469c-bc19-b0afb73a7134-kube-api-access-wvdqm\") pod \"community-operators-t7mjz\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.227651 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.697142 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7mjz"] Oct 07 23:58:28 crc kubenswrapper[4871]: I1007 23:58:28.994079 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbe71c83-937f-4cce-8ea2-26c78aeb84de" path="/var/lib/kubelet/pods/dbe71c83-937f-4cce-8ea2-26c78aeb84de/volumes" Oct 07 23:58:29 crc kubenswrapper[4871]: I1007 23:58:29.145119 4871 generic.go:334] "Generic (PLEG): container finished" podID="caece94b-ab07-469c-bc19-b0afb73a7134" containerID="c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc" exitCode=0 Oct 07 23:58:29 crc kubenswrapper[4871]: I1007 23:58:29.145183 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mjz" event={"ID":"caece94b-ab07-469c-bc19-b0afb73a7134","Type":"ContainerDied","Data":"c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc"} Oct 07 23:58:29 crc kubenswrapper[4871]: I1007 23:58:29.145227 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mjz" event={"ID":"caece94b-ab07-469c-bc19-b0afb73a7134","Type":"ContainerStarted","Data":"f0000e477fd78417c0d8b37f877ed34a3de26c448b4b6108ba493a3473213d47"} Oct 07 23:58:30 crc kubenswrapper[4871]: I1007 23:58:30.175380 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mjz" event={"ID":"caece94b-ab07-469c-bc19-b0afb73a7134","Type":"ContainerStarted","Data":"4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce"} Oct 07 23:58:32 crc kubenswrapper[4871]: I1007 23:58:32.216239 4871 generic.go:334] "Generic (PLEG): container finished" podID="caece94b-ab07-469c-bc19-b0afb73a7134" containerID="4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce" exitCode=0 Oct 07 23:58:32 crc kubenswrapper[4871]: I1007 23:58:32.216312 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mjz" event={"ID":"caece94b-ab07-469c-bc19-b0afb73a7134","Type":"ContainerDied","Data":"4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce"} Oct 07 23:58:32 crc kubenswrapper[4871]: I1007 23:58:32.983062 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:58:32 crc kubenswrapper[4871]: E1007 23:58:32.983989 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:58:33 crc kubenswrapper[4871]: I1007 23:58:33.237492 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mjz" event={"ID":"caece94b-ab07-469c-bc19-b0afb73a7134","Type":"ContainerStarted","Data":"373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce"} Oct 07 23:58:33 crc kubenswrapper[4871]: I1007 23:58:33.256241 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t7mjz" podStartSLOduration=2.647640439 podStartE2EDuration="6.256216899s" podCreationTimestamp="2025-10-07 23:58:27 +0000 UTC" firstStartedPulling="2025-10-07 23:58:29.147538423 +0000 UTC m=+6582.950236506" lastFinishedPulling="2025-10-07 23:58:32.756114863 +0000 UTC m=+6586.558812966" observedRunningTime="2025-10-07 23:58:33.254407332 +0000 UTC m=+6587.057105415" watchObservedRunningTime="2025-10-07 23:58:33.256216899 +0000 UTC m=+6587.058914972" Oct 07 23:58:38 crc kubenswrapper[4871]: I1007 23:58:38.228156 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:38 crc kubenswrapper[4871]: I1007 23:58:38.229899 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:39 crc kubenswrapper[4871]: I1007 23:58:39.316423 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-t7mjz" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" containerName="registry-server" probeResult="failure" output=< Oct 07 23:58:39 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 23:58:39 crc kubenswrapper[4871]: > Oct 07 23:58:45 crc kubenswrapper[4871]: I1007 23:58:45.983153 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:58:45 crc kubenswrapper[4871]: E1007 23:58:45.984081 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:58:48 crc kubenswrapper[4871]: I1007 23:58:48.280080 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:48 crc kubenswrapper[4871]: I1007 23:58:48.348585 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:48 crc kubenswrapper[4871]: I1007 23:58:48.522961 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7mjz"] Oct 07 23:58:49 crc kubenswrapper[4871]: I1007 23:58:49.453655 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t7mjz" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" containerName="registry-server" containerID="cri-o://373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce" gracePeriod=2 Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.063649 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.211319 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-utilities\") pod \"caece94b-ab07-469c-bc19-b0afb73a7134\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.211889 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvdqm\" (UniqueName: \"kubernetes.io/projected/caece94b-ab07-469c-bc19-b0afb73a7134-kube-api-access-wvdqm\") pod \"caece94b-ab07-469c-bc19-b0afb73a7134\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.211974 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-catalog-content\") pod \"caece94b-ab07-469c-bc19-b0afb73a7134\" (UID: \"caece94b-ab07-469c-bc19-b0afb73a7134\") " Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.217898 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-utilities" (OuterVolumeSpecName: "utilities") pod "caece94b-ab07-469c-bc19-b0afb73a7134" (UID: "caece94b-ab07-469c-bc19-b0afb73a7134"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.223847 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caece94b-ab07-469c-bc19-b0afb73a7134-kube-api-access-wvdqm" (OuterVolumeSpecName: "kube-api-access-wvdqm") pod "caece94b-ab07-469c-bc19-b0afb73a7134" (UID: "caece94b-ab07-469c-bc19-b0afb73a7134"). InnerVolumeSpecName "kube-api-access-wvdqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.279748 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "caece94b-ab07-469c-bc19-b0afb73a7134" (UID: "caece94b-ab07-469c-bc19-b0afb73a7134"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.314910 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.314936 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvdqm\" (UniqueName: \"kubernetes.io/projected/caece94b-ab07-469c-bc19-b0afb73a7134-kube-api-access-wvdqm\") on node \"crc\" DevicePath \"\"" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.314945 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caece94b-ab07-469c-bc19-b0afb73a7134-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.468524 4871 generic.go:334] "Generic (PLEG): container finished" podID="caece94b-ab07-469c-bc19-b0afb73a7134" containerID="373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce" exitCode=0 Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.468575 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mjz" event={"ID":"caece94b-ab07-469c-bc19-b0afb73a7134","Type":"ContainerDied","Data":"373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce"} Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.468606 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mjz" event={"ID":"caece94b-ab07-469c-bc19-b0afb73a7134","Type":"ContainerDied","Data":"f0000e477fd78417c0d8b37f877ed34a3de26c448b4b6108ba493a3473213d47"} Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.468625 4871 scope.go:117] "RemoveContainer" containerID="373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.468747 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7mjz" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.505776 4871 scope.go:117] "RemoveContainer" containerID="4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.510913 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7mjz"] Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.529250 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t7mjz"] Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.541379 4871 scope.go:117] "RemoveContainer" containerID="c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.580014 4871 scope.go:117] "RemoveContainer" containerID="373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce" Oct 07 23:58:50 crc kubenswrapper[4871]: E1007 23:58:50.581634 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce\": container with ID starting with 373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce not found: ID does not exist" containerID="373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.581676 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce"} err="failed to get container status \"373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce\": rpc error: code = NotFound desc = could not find container \"373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce\": container with ID starting with 373e2180bf0f159047a22171d3e8ad6ffc2d650c98dad32f36fa4fb2c121fdce not found: ID does not exist" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.581702 4871 scope.go:117] "RemoveContainer" containerID="4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce" Oct 07 23:58:50 crc kubenswrapper[4871]: E1007 23:58:50.582773 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce\": container with ID starting with 4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce not found: ID does not exist" containerID="4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.582825 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce"} err="failed to get container status \"4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce\": rpc error: code = NotFound desc = could not find container \"4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce\": container with ID starting with 4fdd79f77b9ced6034041f52d72e90e9ebe9becb890352e489331d37d3a232ce not found: ID does not exist" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.582846 4871 scope.go:117] "RemoveContainer" containerID="c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc" Oct 07 23:58:50 crc kubenswrapper[4871]: E1007 23:58:50.583627 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc\": container with ID starting with c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc not found: ID does not exist" containerID="c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc" Oct 07 23:58:50 crc kubenswrapper[4871]: I1007 23:58:50.583658 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc"} err="failed to get container status \"c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc\": rpc error: code = NotFound desc = could not find container \"c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc\": container with ID starting with c7475ee39ee7ae360797e664de8130235d45da8b37649b474bea5baaeb4a98fc not found: ID does not exist" Oct 07 23:58:51 crc kubenswrapper[4871]: I1007 23:58:51.001241 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" path="/var/lib/kubelet/pods/caece94b-ab07-469c-bc19-b0afb73a7134/volumes" Oct 07 23:58:58 crc kubenswrapper[4871]: I1007 23:58:58.982527 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:58:58 crc kubenswrapper[4871]: E1007 23:58:58.983319 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:59:00 crc kubenswrapper[4871]: I1007 23:59:00.992073 4871 scope.go:117] "RemoveContainer" containerID="a1b1e2021025c96be9d4900631fc35284c4709777780086d38898e33b6b85fb1" Oct 07 23:59:01 crc kubenswrapper[4871]: I1007 23:59:01.026034 4871 scope.go:117] "RemoveContainer" containerID="b06488eec2fb7f1159bede2eb9d48b3ade5e20c2bc3d2f32a86e5f344eb5facb" Oct 07 23:59:01 crc kubenswrapper[4871]: I1007 23:59:01.080214 4871 scope.go:117] "RemoveContainer" containerID="f14c14f5ade2edcf73b172090cb7b45d53273b48c7614eb77f60a4474337ab0d" Oct 07 23:59:11 crc kubenswrapper[4871]: I1007 23:59:11.982613 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:59:11 crc kubenswrapper[4871]: E1007 23:59:11.983566 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:59:15 crc kubenswrapper[4871]: I1007 23:59:15.055731 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-plj8s"] Oct 07 23:59:15 crc kubenswrapper[4871]: I1007 23:59:15.063164 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-plj8s"] Oct 07 23:59:17 crc kubenswrapper[4871]: I1007 23:59:17.007245 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44085888-b090-4d53-964a-def7da4c1b00" path="/var/lib/kubelet/pods/44085888-b090-4d53-964a-def7da4c1b00/volumes" Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.840269 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d2zgk"] Oct 07 23:59:21 crc kubenswrapper[4871]: E1007 23:59:21.841660 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" containerName="extract-utilities" Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.841677 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" containerName="extract-utilities" Oct 07 23:59:21 crc kubenswrapper[4871]: E1007 23:59:21.841709 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" containerName="registry-server" Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.841717 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" containerName="registry-server" Oct 07 23:59:21 crc kubenswrapper[4871]: E1007 23:59:21.841751 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" containerName="extract-content" Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.841758 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" containerName="extract-content" Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.842026 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="caece94b-ab07-469c-bc19-b0afb73a7134" containerName="registry-server" Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.844860 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.876231 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d2zgk"] Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.937016 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p89tg\" (UniqueName: \"kubernetes.io/projected/6055c874-d393-4c7e-9ddb-83ef69ddb08e-kube-api-access-p89tg\") pod \"redhat-operators-d2zgk\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.937112 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-utilities\") pod \"redhat-operators-d2zgk\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:21 crc kubenswrapper[4871]: I1007 23:59:21.937152 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-catalog-content\") pod \"redhat-operators-d2zgk\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.040050 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p89tg\" (UniqueName: \"kubernetes.io/projected/6055c874-d393-4c7e-9ddb-83ef69ddb08e-kube-api-access-p89tg\") pod \"redhat-operators-d2zgk\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.040548 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-utilities\") pod \"redhat-operators-d2zgk\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.040618 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-catalog-content\") pod \"redhat-operators-d2zgk\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.041100 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-utilities\") pod \"redhat-operators-d2zgk\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.041248 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-catalog-content\") pod \"redhat-operators-d2zgk\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.062437 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p89tg\" (UniqueName: \"kubernetes.io/projected/6055c874-d393-4c7e-9ddb-83ef69ddb08e-kube-api-access-p89tg\") pod \"redhat-operators-d2zgk\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.183550 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.678243 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d2zgk"] Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.913066 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2zgk" event={"ID":"6055c874-d393-4c7e-9ddb-83ef69ddb08e","Type":"ContainerStarted","Data":"042cf38dd89f92f56429b381c52fdebb20b5111ddf249243daede4db4265e2df"} Oct 07 23:59:22 crc kubenswrapper[4871]: I1007 23:59:22.913571 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2zgk" event={"ID":"6055c874-d393-4c7e-9ddb-83ef69ddb08e","Type":"ContainerStarted","Data":"caf04da4300ba211d8b202dc2f6be62c29cd2b35c10247ffa9b36bbaffa7432b"} Oct 07 23:59:23 crc kubenswrapper[4871]: I1007 23:59:23.924459 4871 generic.go:334] "Generic (PLEG): container finished" podID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerID="042cf38dd89f92f56429b381c52fdebb20b5111ddf249243daede4db4265e2df" exitCode=0 Oct 07 23:59:23 crc kubenswrapper[4871]: I1007 23:59:23.924853 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2zgk" event={"ID":"6055c874-d393-4c7e-9ddb-83ef69ddb08e","Type":"ContainerDied","Data":"042cf38dd89f92f56429b381c52fdebb20b5111ddf249243daede4db4265e2df"} Oct 07 23:59:23 crc kubenswrapper[4871]: I1007 23:59:23.927332 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 23:59:24 crc kubenswrapper[4871]: I1007 23:59:24.936714 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2zgk" event={"ID":"6055c874-d393-4c7e-9ddb-83ef69ddb08e","Type":"ContainerStarted","Data":"9dcbfcee077dff83a451ae4283810f0f19e9fbf6d90c826d630387d22cc7b359"} Oct 07 23:59:26 crc kubenswrapper[4871]: I1007 23:59:26.992093 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:59:26 crc kubenswrapper[4871]: E1007 23:59:26.992700 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:59:28 crc kubenswrapper[4871]: I1007 23:59:28.987507 4871 generic.go:334] "Generic (PLEG): container finished" podID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerID="9dcbfcee077dff83a451ae4283810f0f19e9fbf6d90c826d630387d22cc7b359" exitCode=0 Oct 07 23:59:29 crc kubenswrapper[4871]: I1007 23:59:29.008305 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2zgk" event={"ID":"6055c874-d393-4c7e-9ddb-83ef69ddb08e","Type":"ContainerDied","Data":"9dcbfcee077dff83a451ae4283810f0f19e9fbf6d90c826d630387d22cc7b359"} Oct 07 23:59:30 crc kubenswrapper[4871]: I1007 23:59:30.002431 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2zgk" event={"ID":"6055c874-d393-4c7e-9ddb-83ef69ddb08e","Type":"ContainerStarted","Data":"7f6eea7c4e0a143a41f6362051e5d54fde493ccd294027d317c82761c3b7a036"} Oct 07 23:59:30 crc kubenswrapper[4871]: I1007 23:59:30.029656 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d2zgk" podStartSLOduration=3.408503924 podStartE2EDuration="9.029628272s" podCreationTimestamp="2025-10-07 23:59:21 +0000 UTC" firstStartedPulling="2025-10-07 23:59:23.926972503 +0000 UTC m=+6637.729670586" lastFinishedPulling="2025-10-07 23:59:29.548096861 +0000 UTC m=+6643.350794934" observedRunningTime="2025-10-07 23:59:30.021200071 +0000 UTC m=+6643.823898144" watchObservedRunningTime="2025-10-07 23:59:30.029628272 +0000 UTC m=+6643.832326365" Oct 07 23:59:32 crc kubenswrapper[4871]: I1007 23:59:32.184380 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:32 crc kubenswrapper[4871]: I1007 23:59:32.184967 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:33 crc kubenswrapper[4871]: I1007 23:59:33.239949 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d2zgk" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerName="registry-server" probeResult="failure" output=< Oct 07 23:59:33 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 07 23:59:33 crc kubenswrapper[4871]: > Oct 07 23:59:37 crc kubenswrapper[4871]: I1007 23:59:37.981991 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:59:37 crc kubenswrapper[4871]: E1007 23:59:37.982581 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 07 23:59:42 crc kubenswrapper[4871]: I1007 23:59:42.288518 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:42 crc kubenswrapper[4871]: I1007 23:59:42.350278 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:46 crc kubenswrapper[4871]: I1007 23:59:46.612340 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d2zgk"] Oct 07 23:59:46 crc kubenswrapper[4871]: I1007 23:59:46.613218 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d2zgk" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerName="registry-server" containerID="cri-o://7f6eea7c4e0a143a41f6362051e5d54fde493ccd294027d317c82761c3b7a036" gracePeriod=2 Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.177860 4871 generic.go:334] "Generic (PLEG): container finished" podID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerID="7f6eea7c4e0a143a41f6362051e5d54fde493ccd294027d317c82761c3b7a036" exitCode=0 Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.177958 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2zgk" event={"ID":"6055c874-d393-4c7e-9ddb-83ef69ddb08e","Type":"ContainerDied","Data":"7f6eea7c4e0a143a41f6362051e5d54fde493ccd294027d317c82761c3b7a036"} Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.178255 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2zgk" event={"ID":"6055c874-d393-4c7e-9ddb-83ef69ddb08e","Type":"ContainerDied","Data":"caf04da4300ba211d8b202dc2f6be62c29cd2b35c10247ffa9b36bbaffa7432b"} Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.178273 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caf04da4300ba211d8b202dc2f6be62c29cd2b35c10247ffa9b36bbaffa7432b" Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.180859 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.341321 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p89tg\" (UniqueName: \"kubernetes.io/projected/6055c874-d393-4c7e-9ddb-83ef69ddb08e-kube-api-access-p89tg\") pod \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.341365 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-utilities\") pod \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.341432 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-catalog-content\") pod \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\" (UID: \"6055c874-d393-4c7e-9ddb-83ef69ddb08e\") " Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.342433 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-utilities" (OuterVolumeSpecName: "utilities") pod "6055c874-d393-4c7e-9ddb-83ef69ddb08e" (UID: "6055c874-d393-4c7e-9ddb-83ef69ddb08e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.354856 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6055c874-d393-4c7e-9ddb-83ef69ddb08e-kube-api-access-p89tg" (OuterVolumeSpecName: "kube-api-access-p89tg") pod "6055c874-d393-4c7e-9ddb-83ef69ddb08e" (UID: "6055c874-d393-4c7e-9ddb-83ef69ddb08e"). InnerVolumeSpecName "kube-api-access-p89tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.444147 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p89tg\" (UniqueName: \"kubernetes.io/projected/6055c874-d393-4c7e-9ddb-83ef69ddb08e-kube-api-access-p89tg\") on node \"crc\" DevicePath \"\"" Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.444183 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.452377 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6055c874-d393-4c7e-9ddb-83ef69ddb08e" (UID: "6055c874-d393-4c7e-9ddb-83ef69ddb08e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 23:59:47 crc kubenswrapper[4871]: I1007 23:59:47.547288 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6055c874-d393-4c7e-9ddb-83ef69ddb08e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 23:59:48 crc kubenswrapper[4871]: I1007 23:59:48.187225 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2zgk" Oct 07 23:59:48 crc kubenswrapper[4871]: I1007 23:59:48.225268 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d2zgk"] Oct 07 23:59:48 crc kubenswrapper[4871]: I1007 23:59:48.238063 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d2zgk"] Oct 07 23:59:49 crc kubenswrapper[4871]: I1007 23:59:49.003318 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" path="/var/lib/kubelet/pods/6055c874-d393-4c7e-9ddb-83ef69ddb08e/volumes" Oct 07 23:59:51 crc kubenswrapper[4871]: I1007 23:59:51.983293 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 07 23:59:51 crc kubenswrapper[4871]: E1007 23:59:51.985692 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.238859 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29331360-qsgv6"] Oct 08 00:00:00 crc kubenswrapper[4871]: E1008 00:00:00.240524 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerName="extract-utilities" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.240547 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerName="extract-utilities" Oct 08 00:00:00 crc kubenswrapper[4871]: E1008 00:00:00.240613 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerName="extract-content" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.240626 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerName="extract-content" Oct 08 00:00:00 crc kubenswrapper[4871]: E1008 00:00:00.240650 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerName="registry-server" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.240659 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerName="registry-server" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.241261 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="6055c874-d393-4c7e-9ddb-83ef69ddb08e" containerName="registry-server" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.242593 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.250247 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x5bn\" (UniqueName: \"kubernetes.io/projected/d980d5b5-997a-408f-aa2b-cfa54324b062-kube-api-access-9x5bn\") pod \"image-pruner-29331360-qsgv6\" (UID: \"d980d5b5-997a-408f-aa2b-cfa54324b062\") " pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.251372 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d980d5b5-997a-408f-aa2b-cfa54324b062-serviceca\") pod \"image-pruner-29331360-qsgv6\" (UID: \"d980d5b5-997a-408f-aa2b-cfa54324b062\") " pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.252903 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.253277 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.258527 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-purge-29331360-6zsn2"] Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.260560 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.268930 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.276814 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29331360-qsgv6"] Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.288412 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-purge-29331360-d65r4"] Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.289922 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.295130 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.296669 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7"] Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.298083 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.307264 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.307854 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.314078 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-purge-29331360-6zsn2"] Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.331925 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7"] Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.349661 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-purge-29331360-d65r4"] Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353411 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x5bn\" (UniqueName: \"kubernetes.io/projected/d980d5b5-997a-408f-aa2b-cfa54324b062-kube-api-access-9x5bn\") pod \"image-pruner-29331360-qsgv6\" (UID: \"d980d5b5-997a-408f-aa2b-cfa54324b062\") " pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353496 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-scripts\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353542 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz9b2\" (UniqueName: \"kubernetes.io/projected/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-kube-api-access-vz9b2\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353570 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-config-volume\") pod \"collect-profiles-29331360-847m7\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353621 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-combined-ca-bundle\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353672 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-config-data\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353701 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d980d5b5-997a-408f-aa2b-cfa54324b062-serviceca\") pod \"image-pruner-29331360-qsgv6\" (UID: \"d980d5b5-997a-408f-aa2b-cfa54324b062\") " pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353768 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-secret-volume\") pod \"collect-profiles-29331360-847m7\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353943 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-combined-ca-bundle\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.353991 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-config-data\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.354035 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjqnf\" (UniqueName: \"kubernetes.io/projected/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-kube-api-access-gjqnf\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.354068 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctcbm\" (UniqueName: \"kubernetes.io/projected/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-kube-api-access-ctcbm\") pod \"collect-profiles-29331360-847m7\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.354098 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-scripts\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.355001 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d980d5b5-997a-408f-aa2b-cfa54324b062-serviceca\") pod \"image-pruner-29331360-qsgv6\" (UID: \"d980d5b5-997a-408f-aa2b-cfa54324b062\") " pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.380077 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x5bn\" (UniqueName: \"kubernetes.io/projected/d980d5b5-997a-408f-aa2b-cfa54324b062-kube-api-access-9x5bn\") pod \"image-pruner-29331360-qsgv6\" (UID: \"d980d5b5-997a-408f-aa2b-cfa54324b062\") " pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.464982 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-combined-ca-bundle\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465058 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-config-data\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465125 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-secret-volume\") pod \"collect-profiles-29331360-847m7\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465188 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-combined-ca-bundle\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465216 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-config-data\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465250 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjqnf\" (UniqueName: \"kubernetes.io/projected/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-kube-api-access-gjqnf\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465279 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctcbm\" (UniqueName: \"kubernetes.io/projected/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-kube-api-access-ctcbm\") pod \"collect-profiles-29331360-847m7\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465304 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-scripts\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465330 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-scripts\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465358 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz9b2\" (UniqueName: \"kubernetes.io/projected/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-kube-api-access-vz9b2\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.465373 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-config-volume\") pod \"collect-profiles-29331360-847m7\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.466189 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-config-volume\") pod \"collect-profiles-29331360-847m7\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.470881 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-combined-ca-bundle\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.473610 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-secret-volume\") pod \"collect-profiles-29331360-847m7\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.473597 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-scripts\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.473743 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-scripts\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.473885 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-combined-ca-bundle\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.473893 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-config-data\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.474489 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-config-data\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.489706 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctcbm\" (UniqueName: \"kubernetes.io/projected/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-kube-api-access-ctcbm\") pod \"collect-profiles-29331360-847m7\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.497587 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz9b2\" (UniqueName: \"kubernetes.io/projected/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-kube-api-access-vz9b2\") pod \"nova-cell0-db-purge-29331360-6zsn2\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.498180 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjqnf\" (UniqueName: \"kubernetes.io/projected/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-kube-api-access-gjqnf\") pod \"nova-cell1-db-purge-29331360-d65r4\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.614461 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.629818 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.641727 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:00 crc kubenswrapper[4871]: I1008 00:00:00.654314 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.114342 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29331360-qsgv6"] Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.242812 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-purge-29331360-6zsn2"] Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.249428 4871 scope.go:117] "RemoveContainer" containerID="d9085652688471218090da443ee3015394e601b9b013e425fcfe2d9ff96a1e22" Oct 08 00:00:01 crc kubenswrapper[4871]: W1008 00:00:01.251419 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dbba49b_2784_4094_8fa7_2c7ba25d6c0c.slice/crio-c18beada53bd3c5127c2c6ed703038eb97a4054bd77dfce777f4aa54a15705a2 WatchSource:0}: Error finding container c18beada53bd3c5127c2c6ed703038eb97a4054bd77dfce777f4aa54a15705a2: Status 404 returned error can't find the container with id c18beada53bd3c5127c2c6ed703038eb97a4054bd77dfce777f4aa54a15705a2 Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.296299 4871 scope.go:117] "RemoveContainer" containerID="cae258903f12b66d45ac5b69b7b9a291017fce79fdb679330c3e6c17a1a2e33e" Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.345407 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7"] Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.354118 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-purge-29331360-6zsn2" event={"ID":"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c","Type":"ContainerStarted","Data":"c18beada53bd3c5127c2c6ed703038eb97a4054bd77dfce777f4aa54a15705a2"} Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.355419 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-purge-29331360-d65r4"] Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.364193 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29331360-qsgv6" event={"ID":"d980d5b5-997a-408f-aa2b-cfa54324b062","Type":"ContainerStarted","Data":"2ee0b505b781dbf4d116dc11ea4ccbcb4ea9274c634eeb9b45291eea055871c6"} Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.364247 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29331360-qsgv6" event={"ID":"d980d5b5-997a-408f-aa2b-cfa54324b062","Type":"ContainerStarted","Data":"d927382071e91978be81048371b28d184f0b0128dc92d4b54efada4c0b4c97bc"} Oct 08 00:00:01 crc kubenswrapper[4871]: W1008 00:00:01.365404 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcca9d1a_a6c4_47ef_bd7a_5c6d1b149779.slice/crio-fba6b61dc57891d21ed106dc9abccb1bd192f5cc59ef978a3f1febd938b1f0ac WatchSource:0}: Error finding container fba6b61dc57891d21ed106dc9abccb1bd192f5cc59ef978a3f1febd938b1f0ac: Status 404 returned error can't find the container with id fba6b61dc57891d21ed106dc9abccb1bd192f5cc59ef978a3f1febd938b1f0ac Oct 08 00:00:01 crc kubenswrapper[4871]: I1008 00:00:01.392608 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29331360-qsgv6" podStartSLOduration=1.392588837 podStartE2EDuration="1.392588837s" podCreationTimestamp="2025-10-08 00:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:00:01.391499069 +0000 UTC m=+6675.194197142" watchObservedRunningTime="2025-10-08 00:00:01.392588837 +0000 UTC m=+6675.195286910" Oct 08 00:00:02 crc kubenswrapper[4871]: I1008 00:00:02.385088 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-purge-29331360-d65r4" event={"ID":"dcde4412-ee28-45eb-91f1-2718fa9fe0a5","Type":"ContainerStarted","Data":"17e0095608fd3b31c391bdea178b3fa698ac9b2e95732544ed80812934291e95"} Oct 08 00:00:02 crc kubenswrapper[4871]: I1008 00:00:02.385974 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-purge-29331360-d65r4" event={"ID":"dcde4412-ee28-45eb-91f1-2718fa9fe0a5","Type":"ContainerStarted","Data":"89922c9af51bb8a5e5b43206f092c766c7475edd660865818254c27cfd19e178"} Oct 08 00:00:02 crc kubenswrapper[4871]: I1008 00:00:02.389045 4871 generic.go:334] "Generic (PLEG): container finished" podID="bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779" containerID="65c58dd948bc3c17b2b9f07eb8e6e9f603c4b22926aa2f24fced696ab5d766d8" exitCode=0 Oct 08 00:00:02 crc kubenswrapper[4871]: I1008 00:00:02.389188 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" event={"ID":"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779","Type":"ContainerDied","Data":"65c58dd948bc3c17b2b9f07eb8e6e9f603c4b22926aa2f24fced696ab5d766d8"} Oct 08 00:00:02 crc kubenswrapper[4871]: I1008 00:00:02.389212 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" event={"ID":"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779","Type":"ContainerStarted","Data":"fba6b61dc57891d21ed106dc9abccb1bd192f5cc59ef978a3f1febd938b1f0ac"} Oct 08 00:00:02 crc kubenswrapper[4871]: I1008 00:00:02.398062 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-purge-29331360-6zsn2" event={"ID":"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c","Type":"ContainerStarted","Data":"6a3de0351159336e3f78a63a0f11205385842494b851bd96ea10b72d4893e451"} Oct 08 00:00:02 crc kubenswrapper[4871]: I1008 00:00:02.403007 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-purge-29331360-d65r4" podStartSLOduration=2.402993399 podStartE2EDuration="2.402993399s" podCreationTimestamp="2025-10-08 00:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:00:02.39845366 +0000 UTC m=+6676.201151733" watchObservedRunningTime="2025-10-08 00:00:02.402993399 +0000 UTC m=+6676.205691472" Oct 08 00:00:02 crc kubenswrapper[4871]: I1008 00:00:02.425404 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-purge-29331360-6zsn2" podStartSLOduration=2.425386856 podStartE2EDuration="2.425386856s" podCreationTimestamp="2025-10-08 00:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:00:02.415221609 +0000 UTC m=+6676.217919682" watchObservedRunningTime="2025-10-08 00:00:02.425386856 +0000 UTC m=+6676.228084929" Oct 08 00:00:03 crc kubenswrapper[4871]: I1008 00:00:03.415638 4871 generic.go:334] "Generic (PLEG): container finished" podID="d980d5b5-997a-408f-aa2b-cfa54324b062" containerID="2ee0b505b781dbf4d116dc11ea4ccbcb4ea9274c634eeb9b45291eea055871c6" exitCode=0 Oct 08 00:00:03 crc kubenswrapper[4871]: I1008 00:00:03.415893 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29331360-qsgv6" event={"ID":"d980d5b5-997a-408f-aa2b-cfa54324b062","Type":"ContainerDied","Data":"2ee0b505b781dbf4d116dc11ea4ccbcb4ea9274c634eeb9b45291eea055871c6"} Oct 08 00:00:03 crc kubenswrapper[4871]: I1008 00:00:03.830492 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:03 crc kubenswrapper[4871]: I1008 00:00:03.964887 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctcbm\" (UniqueName: \"kubernetes.io/projected/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-kube-api-access-ctcbm\") pod \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " Oct 08 00:00:03 crc kubenswrapper[4871]: I1008 00:00:03.965125 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-secret-volume\") pod \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " Oct 08 00:00:03 crc kubenswrapper[4871]: I1008 00:00:03.965565 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-config-volume\") pod \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\" (UID: \"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779\") " Oct 08 00:00:03 crc kubenswrapper[4871]: I1008 00:00:03.966618 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-config-volume" (OuterVolumeSpecName: "config-volume") pod "bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779" (UID: "bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:00:03 crc kubenswrapper[4871]: I1008 00:00:03.970830 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779" (UID: "bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:00:03 crc kubenswrapper[4871]: I1008 00:00:03.971884 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-kube-api-access-ctcbm" (OuterVolumeSpecName: "kube-api-access-ctcbm") pod "bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779" (UID: "bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779"). InnerVolumeSpecName "kube-api-access-ctcbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.068305 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.068346 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctcbm\" (UniqueName: \"kubernetes.io/projected/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-kube-api-access-ctcbm\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.068359 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.432172 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" event={"ID":"bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779","Type":"ContainerDied","Data":"fba6b61dc57891d21ed106dc9abccb1bd192f5cc59ef978a3f1febd938b1f0ac"} Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.432287 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fba6b61dc57891d21ed106dc9abccb1bd192f5cc59ef978a3f1febd938b1f0ac" Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.432205 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7" Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.798480 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.902068 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9"] Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.911323 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331315-76nm9"] Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.991098 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 08 00:00:04 crc kubenswrapper[4871]: E1008 00:00:04.991372 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.992263 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x5bn\" (UniqueName: \"kubernetes.io/projected/d980d5b5-997a-408f-aa2b-cfa54324b062-kube-api-access-9x5bn\") pod \"d980d5b5-997a-408f-aa2b-cfa54324b062\" (UID: \"d980d5b5-997a-408f-aa2b-cfa54324b062\") " Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.992305 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d980d5b5-997a-408f-aa2b-cfa54324b062-serviceca\") pod \"d980d5b5-997a-408f-aa2b-cfa54324b062\" (UID: \"d980d5b5-997a-408f-aa2b-cfa54324b062\") " Oct 08 00:00:04 crc kubenswrapper[4871]: I1008 00:00:04.993088 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d980d5b5-997a-408f-aa2b-cfa54324b062-serviceca" (OuterVolumeSpecName: "serviceca") pod "d980d5b5-997a-408f-aa2b-cfa54324b062" (UID: "d980d5b5-997a-408f-aa2b-cfa54324b062"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:00:05 crc kubenswrapper[4871]: I1008 00:00:05.000372 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d980d5b5-997a-408f-aa2b-cfa54324b062-kube-api-access-9x5bn" (OuterVolumeSpecName: "kube-api-access-9x5bn") pod "d980d5b5-997a-408f-aa2b-cfa54324b062" (UID: "d980d5b5-997a-408f-aa2b-cfa54324b062"). InnerVolumeSpecName "kube-api-access-9x5bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:00:05 crc kubenswrapper[4871]: I1008 00:00:05.000747 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d395bf3-85d2-4f95-af96-d62c4bc13ab0" path="/var/lib/kubelet/pods/2d395bf3-85d2-4f95-af96-d62c4bc13ab0/volumes" Oct 08 00:00:05 crc kubenswrapper[4871]: I1008 00:00:05.094962 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x5bn\" (UniqueName: \"kubernetes.io/projected/d980d5b5-997a-408f-aa2b-cfa54324b062-kube-api-access-9x5bn\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:05 crc kubenswrapper[4871]: I1008 00:00:05.094992 4871 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d980d5b5-997a-408f-aa2b-cfa54324b062-serviceca\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:05 crc kubenswrapper[4871]: I1008 00:00:05.442148 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29331360-qsgv6" event={"ID":"d980d5b5-997a-408f-aa2b-cfa54324b062","Type":"ContainerDied","Data":"d927382071e91978be81048371b28d184f0b0128dc92d4b54efada4c0b4c97bc"} Oct 08 00:00:05 crc kubenswrapper[4871]: I1008 00:00:05.442418 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d927382071e91978be81048371b28d184f0b0128dc92d4b54efada4c0b4c97bc" Oct 08 00:00:05 crc kubenswrapper[4871]: I1008 00:00:05.442202 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29331360-qsgv6" Oct 08 00:00:08 crc kubenswrapper[4871]: I1008 00:00:08.483057 4871 generic.go:334] "Generic (PLEG): container finished" podID="dcde4412-ee28-45eb-91f1-2718fa9fe0a5" containerID="17e0095608fd3b31c391bdea178b3fa698ac9b2e95732544ed80812934291e95" exitCode=0 Oct 08 00:00:08 crc kubenswrapper[4871]: I1008 00:00:08.483167 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-purge-29331360-d65r4" event={"ID":"dcde4412-ee28-45eb-91f1-2718fa9fe0a5","Type":"ContainerDied","Data":"17e0095608fd3b31c391bdea178b3fa698ac9b2e95732544ed80812934291e95"} Oct 08 00:00:08 crc kubenswrapper[4871]: I1008 00:00:08.488852 4871 generic.go:334] "Generic (PLEG): container finished" podID="8dbba49b-2784-4094-8fa7-2c7ba25d6c0c" containerID="6a3de0351159336e3f78a63a0f11205385842494b851bd96ea10b72d4893e451" exitCode=0 Oct 08 00:00:08 crc kubenswrapper[4871]: I1008 00:00:08.489020 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-purge-29331360-6zsn2" event={"ID":"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c","Type":"ContainerDied","Data":"6a3de0351159336e3f78a63a0f11205385842494b851bd96ea10b72d4893e451"} Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.031868 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.044871 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.218167 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjqnf\" (UniqueName: \"kubernetes.io/projected/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-kube-api-access-gjqnf\") pod \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.218316 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-config-data\") pod \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.218411 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-combined-ca-bundle\") pod \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.218435 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-combined-ca-bundle\") pod \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.219278 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-scripts\") pod \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.219437 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-scripts\") pod \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.219515 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz9b2\" (UniqueName: \"kubernetes.io/projected/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-kube-api-access-vz9b2\") pod \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\" (UID: \"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c\") " Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.219551 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-config-data\") pod \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\" (UID: \"dcde4412-ee28-45eb-91f1-2718fa9fe0a5\") " Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.225079 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-scripts" (OuterVolumeSpecName: "scripts") pod "dcde4412-ee28-45eb-91f1-2718fa9fe0a5" (UID: "dcde4412-ee28-45eb-91f1-2718fa9fe0a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.226552 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-kube-api-access-gjqnf" (OuterVolumeSpecName: "kube-api-access-gjqnf") pod "dcde4412-ee28-45eb-91f1-2718fa9fe0a5" (UID: "dcde4412-ee28-45eb-91f1-2718fa9fe0a5"). InnerVolumeSpecName "kube-api-access-gjqnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.227019 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-kube-api-access-vz9b2" (OuterVolumeSpecName: "kube-api-access-vz9b2") pod "8dbba49b-2784-4094-8fa7-2c7ba25d6c0c" (UID: "8dbba49b-2784-4094-8fa7-2c7ba25d6c0c"). InnerVolumeSpecName "kube-api-access-vz9b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.235942 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-scripts" (OuterVolumeSpecName: "scripts") pod "8dbba49b-2784-4094-8fa7-2c7ba25d6c0c" (UID: "8dbba49b-2784-4094-8fa7-2c7ba25d6c0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.250707 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dcde4412-ee28-45eb-91f1-2718fa9fe0a5" (UID: "dcde4412-ee28-45eb-91f1-2718fa9fe0a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.251898 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8dbba49b-2784-4094-8fa7-2c7ba25d6c0c" (UID: "8dbba49b-2784-4094-8fa7-2c7ba25d6c0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.252997 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-config-data" (OuterVolumeSpecName: "config-data") pod "8dbba49b-2784-4094-8fa7-2c7ba25d6c0c" (UID: "8dbba49b-2784-4094-8fa7-2c7ba25d6c0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.263968 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-config-data" (OuterVolumeSpecName: "config-data") pod "dcde4412-ee28-45eb-91f1-2718fa9fe0a5" (UID: "dcde4412-ee28-45eb-91f1-2718fa9fe0a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.322971 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.323026 4871 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.323044 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz9b2\" (UniqueName: \"kubernetes.io/projected/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-kube-api-access-vz9b2\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.323064 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.323080 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjqnf\" (UniqueName: \"kubernetes.io/projected/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-kube-api-access-gjqnf\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.323096 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.323111 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcde4412-ee28-45eb-91f1-2718fa9fe0a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.323129 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbba49b-2784-4094-8fa7-2c7ba25d6c0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.512415 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-purge-29331360-6zsn2" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.512488 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-purge-29331360-6zsn2" event={"ID":"8dbba49b-2784-4094-8fa7-2c7ba25d6c0c","Type":"ContainerDied","Data":"c18beada53bd3c5127c2c6ed703038eb97a4054bd77dfce777f4aa54a15705a2"} Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.512589 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c18beada53bd3c5127c2c6ed703038eb97a4054bd77dfce777f4aa54a15705a2" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.515728 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-purge-29331360-d65r4" event={"ID":"dcde4412-ee28-45eb-91f1-2718fa9fe0a5","Type":"ContainerDied","Data":"89922c9af51bb8a5e5b43206f092c766c7475edd660865818254c27cfd19e178"} Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.515788 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89922c9af51bb8a5e5b43206f092c766c7475edd660865818254c27cfd19e178" Oct 08 00:00:10 crc kubenswrapper[4871]: I1008 00:00:10.515867 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-purge-29331360-d65r4" Oct 08 00:00:18 crc kubenswrapper[4871]: I1008 00:00:18.983766 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 08 00:00:19 crc kubenswrapper[4871]: I1008 00:00:19.622148 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"12ef4c6e132b32fbb6956c2b731cbfff15694db1157433b3657f2a0b245e4289"} Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.181765 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-purge-29331361-xd69z"] Oct 08 00:01:00 crc kubenswrapper[4871]: E1008 00:01:00.183225 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcde4412-ee28-45eb-91f1-2718fa9fe0a5" containerName="nova-manage" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.183243 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcde4412-ee28-45eb-91f1-2718fa9fe0a5" containerName="nova-manage" Oct 08 00:01:00 crc kubenswrapper[4871]: E1008 00:01:00.183291 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d980d5b5-997a-408f-aa2b-cfa54324b062" containerName="image-pruner" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.183300 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d980d5b5-997a-408f-aa2b-cfa54324b062" containerName="image-pruner" Oct 08 00:01:00 crc kubenswrapper[4871]: E1008 00:01:00.183311 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbba49b-2784-4094-8fa7-2c7ba25d6c0c" containerName="nova-manage" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.183320 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbba49b-2784-4094-8fa7-2c7ba25d6c0c" containerName="nova-manage" Oct 08 00:01:00 crc kubenswrapper[4871]: E1008 00:01:00.183349 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779" containerName="collect-profiles" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.183359 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779" containerName="collect-profiles" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.183783 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbba49b-2784-4094-8fa7-2c7ba25d6c0c" containerName="nova-manage" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.183824 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779" containerName="collect-profiles" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.183842 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcde4412-ee28-45eb-91f1-2718fa9fe0a5" containerName="nova-manage" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.183859 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d980d5b5-997a-408f-aa2b-cfa54324b062" containerName="image-pruner" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.184910 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.210747 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-purge-29331361-xd69z"] Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.220530 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29331361-9wnxt"] Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.222340 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.239072 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-purge-29331361-crl8s"] Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.240940 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.252412 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-purge-29331361-crl8s"] Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.272160 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-purge-29331361-shv9p"] Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.274323 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.277217 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.292688 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29331361-9wnxt"] Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.313193 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-purge-29331361-shv9p"] Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.314785 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-combined-ca-bundle\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.314940 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7lpt\" (UniqueName: \"kubernetes.io/projected/6dacc481-da68-45bf-b7a2-24f796328ee1-kube-api-access-v7lpt\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.315235 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-db-purge-config-data\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.315341 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-config-data\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.416818 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn4pc\" (UniqueName: \"kubernetes.io/projected/35542928-206a-4b3f-bf2e-ba4ea4449cc3-kube-api-access-mn4pc\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.416888 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-db-purge-config-data\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.416923 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-fernet-keys\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.416959 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjzbs\" (UniqueName: \"kubernetes.io/projected/401dc5ea-abe5-43e0-9d3f-095574675437-kube-api-access-tjzbs\") pod \"manila-db-purge-29331361-crl8s\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417039 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-combined-ca-bundle\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417111 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-db-purge-config-data\") pod \"manila-db-purge-29331361-crl8s\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417145 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-combined-ca-bundle\") pod \"manila-db-purge-29331361-crl8s\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417172 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-combined-ca-bundle\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417214 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-config-data\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417250 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7lpt\" (UniqueName: \"kubernetes.io/projected/6dacc481-da68-45bf-b7a2-24f796328ee1-kube-api-access-v7lpt\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417301 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-config-data\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417380 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/80141182-14e1-4535-a176-4c69d98eaaa5-kube-api-access-ctpnw\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417506 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-db-purge-config-data\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417543 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-combined-ca-bundle\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.417617 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-config-data\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.428578 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-combined-ca-bundle\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.429596 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-config-data\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.430746 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-db-purge-config-data\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.437232 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7lpt\" (UniqueName: \"kubernetes.io/projected/6dacc481-da68-45bf-b7a2-24f796328ee1-kube-api-access-v7lpt\") pod \"cinder-db-purge-29331361-xd69z\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.519357 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-combined-ca-bundle\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.519514 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.519571 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn4pc\" (UniqueName: \"kubernetes.io/projected/35542928-206a-4b3f-bf2e-ba4ea4449cc3-kube-api-access-mn4pc\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.519635 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-db-purge-config-data\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.519684 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-fernet-keys\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.519759 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjzbs\" (UniqueName: \"kubernetes.io/projected/401dc5ea-abe5-43e0-9d3f-095574675437-kube-api-access-tjzbs\") pod \"manila-db-purge-29331361-crl8s\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.519847 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-combined-ca-bundle\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.519905 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-db-purge-config-data\") pod \"manila-db-purge-29331361-crl8s\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.520386 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-combined-ca-bundle\") pod \"manila-db-purge-29331361-crl8s\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.520448 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-config-data\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.520538 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-config-data\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.520625 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/80141182-14e1-4535-a176-4c69d98eaaa5-kube-api-access-ctpnw\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.527642 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-combined-ca-bundle\") pod \"manila-db-purge-29331361-crl8s\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.529722 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-config-data\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.535552 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/80141182-14e1-4535-a176-4c69d98eaaa5-kube-api-access-ctpnw\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.536418 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjzbs\" (UniqueName: \"kubernetes.io/projected/401dc5ea-abe5-43e0-9d3f-095574675437-kube-api-access-tjzbs\") pod \"manila-db-purge-29331361-crl8s\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.536974 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn4pc\" (UniqueName: \"kubernetes.io/projected/35542928-206a-4b3f-bf2e-ba4ea4449cc3-kube-api-access-mn4pc\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.537538 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-db-purge-config-data\") pod \"manila-db-purge-29331361-crl8s\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.538095 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-db-purge-config-data\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.538444 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-fernet-keys\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.538484 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-combined-ca-bundle\") pod \"glance-db-purge-29331361-shv9p\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.545220 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-config-data\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.545964 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-combined-ca-bundle\") pod \"keystone-cron-29331361-9wnxt\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.549621 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.586352 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:00 crc kubenswrapper[4871]: I1008 00:01:00.599872 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:01 crc kubenswrapper[4871]: I1008 00:01:01.232879 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29331361-9wnxt"] Oct 08 00:01:01 crc kubenswrapper[4871]: W1008 00:01:01.236979 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35542928_206a_4b3f_bf2e_ba4ea4449cc3.slice/crio-08714cde88232c0eb877a1d133dbb78216bad3596d0425bb41c65932590ee510 WatchSource:0}: Error finding container 08714cde88232c0eb877a1d133dbb78216bad3596d0425bb41c65932590ee510: Status 404 returned error can't find the container with id 08714cde88232c0eb877a1d133dbb78216bad3596d0425bb41c65932590ee510 Oct 08 00:01:01 crc kubenswrapper[4871]: I1008 00:01:01.240362 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-purge-29331361-xd69z"] Oct 08 00:01:01 crc kubenswrapper[4871]: I1008 00:01:01.373429 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-purge-29331361-shv9p"] Oct 08 00:01:01 crc kubenswrapper[4871]: I1008 00:01:01.438552 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-purge-29331361-crl8s"] Oct 08 00:01:01 crc kubenswrapper[4871]: I1008 00:01:01.440517 4871 scope.go:117] "RemoveContainer" containerID="f6e7c38f0f0b0e1e2e7b44d963307c35a3e49bed289b3b56d39cd1c6213331b0" Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.183253 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-purge-29331361-xd69z" event={"ID":"6dacc481-da68-45bf-b7a2-24f796328ee1","Type":"ContainerStarted","Data":"4e4076a9fe17c12958f250974472c142928b5439f7efcb4ce628dc5665690531"} Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.183596 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-purge-29331361-xd69z" event={"ID":"6dacc481-da68-45bf-b7a2-24f796328ee1","Type":"ContainerStarted","Data":"04f7d8b72f555524da5e31144f107690d6a8949774b3d07f5940eb674d575dee"} Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.185309 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-purge-29331361-crl8s" event={"ID":"401dc5ea-abe5-43e0-9d3f-095574675437","Type":"ContainerStarted","Data":"4472655f62f893f03fd329c11cf7803f0d84ca418e7bdd2884a3adabd560f21d"} Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.185353 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-purge-29331361-crl8s" event={"ID":"401dc5ea-abe5-43e0-9d3f-095574675437","Type":"ContainerStarted","Data":"bd71f50d12d5d8d6d581173d0a992f8a01f1b31a1b353908e5f2809e60ce33aa"} Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.188139 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-purge-29331361-shv9p" event={"ID":"80141182-14e1-4535-a176-4c69d98eaaa5","Type":"ContainerStarted","Data":"af748d02ecc27716bc73d791d79db7c1c5820bc2d2b9d4b6b9b11f7b54c555fb"} Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.189094 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-purge-29331361-shv9p" event={"ID":"80141182-14e1-4535-a176-4c69d98eaaa5","Type":"ContainerStarted","Data":"e96e90469aa5c04af05bd4058011f7b93cac98768de3b0d48ab7e3c7164bbe8f"} Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.192117 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331361-9wnxt" event={"ID":"35542928-206a-4b3f-bf2e-ba4ea4449cc3","Type":"ContainerStarted","Data":"df0ba2521bd917784d57c08cc2a77f3ce6be40d2d7696b1ad4a9e58fbb2f1182"} Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.192164 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331361-9wnxt" event={"ID":"35542928-206a-4b3f-bf2e-ba4ea4449cc3","Type":"ContainerStarted","Data":"08714cde88232c0eb877a1d133dbb78216bad3596d0425bb41c65932590ee510"} Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.210694 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-purge-29331361-xd69z" podStartSLOduration=2.210665717 podStartE2EDuration="2.210665717s" podCreationTimestamp="2025-10-08 00:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:01:02.198405355 +0000 UTC m=+6736.001103448" watchObservedRunningTime="2025-10-08 00:01:02.210665717 +0000 UTC m=+6736.013363830" Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.228529 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-purge-29331361-shv9p" podStartSLOduration=2.228474114 podStartE2EDuration="2.228474114s" podCreationTimestamp="2025-10-08 00:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:01:02.214892878 +0000 UTC m=+6736.017590951" watchObservedRunningTime="2025-10-08 00:01:02.228474114 +0000 UTC m=+6736.031172207" Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.247150 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29331361-9wnxt" podStartSLOduration=2.247126653 podStartE2EDuration="2.247126653s" podCreationTimestamp="2025-10-08 00:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:01:02.234503682 +0000 UTC m=+6736.037201765" watchObservedRunningTime="2025-10-08 00:01:02.247126653 +0000 UTC m=+6736.049824736" Oct 08 00:01:02 crc kubenswrapper[4871]: I1008 00:01:02.264314 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-purge-29331361-crl8s" podStartSLOduration=2.264294403 podStartE2EDuration="2.264294403s" podCreationTimestamp="2025-10-08 00:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:01:02.253903791 +0000 UTC m=+6736.056601864" watchObservedRunningTime="2025-10-08 00:01:02.264294403 +0000 UTC m=+6736.066992476" Oct 08 00:01:04 crc kubenswrapper[4871]: I1008 00:01:04.219054 4871 generic.go:334] "Generic (PLEG): container finished" podID="35542928-206a-4b3f-bf2e-ba4ea4449cc3" containerID="df0ba2521bd917784d57c08cc2a77f3ce6be40d2d7696b1ad4a9e58fbb2f1182" exitCode=0 Oct 08 00:01:04 crc kubenswrapper[4871]: I1008 00:01:04.219145 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331361-9wnxt" event={"ID":"35542928-206a-4b3f-bf2e-ba4ea4449cc3","Type":"ContainerDied","Data":"df0ba2521bd917784d57c08cc2a77f3ce6be40d2d7696b1ad4a9e58fbb2f1182"} Oct 08 00:01:04 crc kubenswrapper[4871]: I1008 00:01:04.223601 4871 generic.go:334] "Generic (PLEG): container finished" podID="401dc5ea-abe5-43e0-9d3f-095574675437" containerID="4472655f62f893f03fd329c11cf7803f0d84ca418e7bdd2884a3adabd560f21d" exitCode=0 Oct 08 00:01:04 crc kubenswrapper[4871]: I1008 00:01:04.223710 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-purge-29331361-crl8s" event={"ID":"401dc5ea-abe5-43e0-9d3f-095574675437","Type":"ContainerDied","Data":"4472655f62f893f03fd329c11cf7803f0d84ca418e7bdd2884a3adabd560f21d"} Oct 08 00:01:04 crc kubenswrapper[4871]: I1008 00:01:04.228293 4871 generic.go:334] "Generic (PLEG): container finished" podID="80141182-14e1-4535-a176-4c69d98eaaa5" containerID="af748d02ecc27716bc73d791d79db7c1c5820bc2d2b9d4b6b9b11f7b54c555fb" exitCode=0 Oct 08 00:01:04 crc kubenswrapper[4871]: I1008 00:01:04.228351 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-purge-29331361-shv9p" event={"ID":"80141182-14e1-4535-a176-4c69d98eaaa5","Type":"ContainerDied","Data":"af748d02ecc27716bc73d791d79db7c1c5820bc2d2b9d4b6b9b11f7b54c555fb"} Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.243408 4871 generic.go:334] "Generic (PLEG): container finished" podID="6dacc481-da68-45bf-b7a2-24f796328ee1" containerID="4e4076a9fe17c12958f250974472c142928b5439f7efcb4ce628dc5665690531" exitCode=0 Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.243531 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-purge-29331361-xd69z" event={"ID":"6dacc481-da68-45bf-b7a2-24f796328ee1","Type":"ContainerDied","Data":"4e4076a9fe17c12958f250974472c142928b5439f7efcb4ce628dc5665690531"} Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.815728 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.822096 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.833274 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989288 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-combined-ca-bundle\") pod \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989408 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-db-purge-config-data\") pod \"401dc5ea-abe5-43e0-9d3f-095574675437\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989474 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-fernet-keys\") pod \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989521 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/80141182-14e1-4535-a176-4c69d98eaaa5-kube-api-access-ctpnw\") pod \"80141182-14e1-4535-a176-4c69d98eaaa5\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989572 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-combined-ca-bundle\") pod \"401dc5ea-abe5-43e0-9d3f-095574675437\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989626 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjzbs\" (UniqueName: \"kubernetes.io/projected/401dc5ea-abe5-43e0-9d3f-095574675437-kube-api-access-tjzbs\") pod \"401dc5ea-abe5-43e0-9d3f-095574675437\" (UID: \"401dc5ea-abe5-43e0-9d3f-095574675437\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989641 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-config-data\") pod \"80141182-14e1-4535-a176-4c69d98eaaa5\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989682 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn4pc\" (UniqueName: \"kubernetes.io/projected/35542928-206a-4b3f-bf2e-ba4ea4449cc3-kube-api-access-mn4pc\") pod \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989716 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-config-data\") pod \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\" (UID: \"35542928-206a-4b3f-bf2e-ba4ea4449cc3\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989742 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-db-purge-config-data\") pod \"80141182-14e1-4535-a176-4c69d98eaaa5\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.989806 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-combined-ca-bundle\") pod \"80141182-14e1-4535-a176-4c69d98eaaa5\" (UID: \"80141182-14e1-4535-a176-4c69d98eaaa5\") " Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.997113 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401dc5ea-abe5-43e0-9d3f-095574675437-kube-api-access-tjzbs" (OuterVolumeSpecName: "kube-api-access-tjzbs") pod "401dc5ea-abe5-43e0-9d3f-095574675437" (UID: "401dc5ea-abe5-43e0-9d3f-095574675437"). InnerVolumeSpecName "kube-api-access-tjzbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.997686 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80141182-14e1-4535-a176-4c69d98eaaa5-kube-api-access-ctpnw" (OuterVolumeSpecName: "kube-api-access-ctpnw") pod "80141182-14e1-4535-a176-4c69d98eaaa5" (UID: "80141182-14e1-4535-a176-4c69d98eaaa5"). InnerVolumeSpecName "kube-api-access-ctpnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.998077 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-db-purge-config-data" (OuterVolumeSpecName: "db-purge-config-data") pod "401dc5ea-abe5-43e0-9d3f-095574675437" (UID: "401dc5ea-abe5-43e0-9d3f-095574675437"). InnerVolumeSpecName "db-purge-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.998601 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35542928-206a-4b3f-bf2e-ba4ea4449cc3-kube-api-access-mn4pc" (OuterVolumeSpecName: "kube-api-access-mn4pc") pod "35542928-206a-4b3f-bf2e-ba4ea4449cc3" (UID: "35542928-206a-4b3f-bf2e-ba4ea4449cc3"). InnerVolumeSpecName "kube-api-access-mn4pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:01:05 crc kubenswrapper[4871]: I1008 00:01:05.999195 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "35542928-206a-4b3f-bf2e-ba4ea4449cc3" (UID: "35542928-206a-4b3f-bf2e-ba4ea4449cc3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.002327 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-db-purge-config-data" (OuterVolumeSpecName: "db-purge-config-data") pod "80141182-14e1-4535-a176-4c69d98eaaa5" (UID: "80141182-14e1-4535-a176-4c69d98eaaa5"). InnerVolumeSpecName "db-purge-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.021394 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80141182-14e1-4535-a176-4c69d98eaaa5" (UID: "80141182-14e1-4535-a176-4c69d98eaaa5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.027771 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-config-data" (OuterVolumeSpecName: "config-data") pod "80141182-14e1-4535-a176-4c69d98eaaa5" (UID: "80141182-14e1-4535-a176-4c69d98eaaa5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.028873 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35542928-206a-4b3f-bf2e-ba4ea4449cc3" (UID: "35542928-206a-4b3f-bf2e-ba4ea4449cc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.030740 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "401dc5ea-abe5-43e0-9d3f-095574675437" (UID: "401dc5ea-abe5-43e0-9d3f-095574675437"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.053866 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-config-data" (OuterVolumeSpecName: "config-data") pod "35542928-206a-4b3f-bf2e-ba4ea4449cc3" (UID: "35542928-206a-4b3f-bf2e-ba4ea4449cc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.093948 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094000 4871 reconciler_common.go:293] "Volume detached for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-db-purge-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094021 4871 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094040 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/80141182-14e1-4535-a176-4c69d98eaaa5-kube-api-access-ctpnw\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094059 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401dc5ea-abe5-43e0-9d3f-095574675437-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094075 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjzbs\" (UniqueName: \"kubernetes.io/projected/401dc5ea-abe5-43e0-9d3f-095574675437-kube-api-access-tjzbs\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094092 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094109 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn4pc\" (UniqueName: \"kubernetes.io/projected/35542928-206a-4b3f-bf2e-ba4ea4449cc3-kube-api-access-mn4pc\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094126 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35542928-206a-4b3f-bf2e-ba4ea4449cc3-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094142 4871 reconciler_common.go:293] "Volume detached for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-db-purge-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.094161 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80141182-14e1-4535-a176-4c69d98eaaa5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.257303 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-purge-29331361-shv9p" event={"ID":"80141182-14e1-4535-a176-4c69d98eaaa5","Type":"ContainerDied","Data":"e96e90469aa5c04af05bd4058011f7b93cac98768de3b0d48ab7e3c7164bbe8f"} Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.257703 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e96e90469aa5c04af05bd4058011f7b93cac98768de3b0d48ab7e3c7164bbe8f" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.257379 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-purge-29331361-shv9p" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.259115 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331361-9wnxt" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.259118 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331361-9wnxt" event={"ID":"35542928-206a-4b3f-bf2e-ba4ea4449cc3","Type":"ContainerDied","Data":"08714cde88232c0eb877a1d133dbb78216bad3596d0425bb41c65932590ee510"} Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.259360 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08714cde88232c0eb877a1d133dbb78216bad3596d0425bb41c65932590ee510" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.261925 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-purge-29331361-crl8s" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.262075 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-purge-29331361-crl8s" event={"ID":"401dc5ea-abe5-43e0-9d3f-095574675437","Type":"ContainerDied","Data":"bd71f50d12d5d8d6d581173d0a992f8a01f1b31a1b353908e5f2809e60ce33aa"} Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.262131 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd71f50d12d5d8d6d581173d0a992f8a01f1b31a1b353908e5f2809e60ce33aa" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.785493 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.913877 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-db-purge-config-data\") pod \"6dacc481-da68-45bf-b7a2-24f796328ee1\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.913938 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7lpt\" (UniqueName: \"kubernetes.io/projected/6dacc481-da68-45bf-b7a2-24f796328ee1-kube-api-access-v7lpt\") pod \"6dacc481-da68-45bf-b7a2-24f796328ee1\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.914045 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-config-data\") pod \"6dacc481-da68-45bf-b7a2-24f796328ee1\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.914117 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-combined-ca-bundle\") pod \"6dacc481-da68-45bf-b7a2-24f796328ee1\" (UID: \"6dacc481-da68-45bf-b7a2-24f796328ee1\") " Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.919458 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-db-purge-config-data" (OuterVolumeSpecName: "db-purge-config-data") pod "6dacc481-da68-45bf-b7a2-24f796328ee1" (UID: "6dacc481-da68-45bf-b7a2-24f796328ee1"). InnerVolumeSpecName "db-purge-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.938308 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dacc481-da68-45bf-b7a2-24f796328ee1-kube-api-access-v7lpt" (OuterVolumeSpecName: "kube-api-access-v7lpt") pod "6dacc481-da68-45bf-b7a2-24f796328ee1" (UID: "6dacc481-da68-45bf-b7a2-24f796328ee1"). InnerVolumeSpecName "kube-api-access-v7lpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.948060 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-config-data" (OuterVolumeSpecName: "config-data") pod "6dacc481-da68-45bf-b7a2-24f796328ee1" (UID: "6dacc481-da68-45bf-b7a2-24f796328ee1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:06 crc kubenswrapper[4871]: I1008 00:01:06.966126 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6dacc481-da68-45bf-b7a2-24f796328ee1" (UID: "6dacc481-da68-45bf-b7a2-24f796328ee1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:01:07 crc kubenswrapper[4871]: I1008 00:01:07.020131 4871 reconciler_common.go:293] "Volume detached for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-db-purge-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:07 crc kubenswrapper[4871]: I1008 00:01:07.020174 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7lpt\" (UniqueName: \"kubernetes.io/projected/6dacc481-da68-45bf-b7a2-24f796328ee1-kube-api-access-v7lpt\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:07 crc kubenswrapper[4871]: I1008 00:01:07.020190 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:07 crc kubenswrapper[4871]: I1008 00:01:07.020202 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dacc481-da68-45bf-b7a2-24f796328ee1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:07 crc kubenswrapper[4871]: I1008 00:01:07.273676 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-purge-29331361-xd69z" event={"ID":"6dacc481-da68-45bf-b7a2-24f796328ee1","Type":"ContainerDied","Data":"04f7d8b72f555524da5e31144f107690d6a8949774b3d07f5940eb674d575dee"} Oct 08 00:01:07 crc kubenswrapper[4871]: I1008 00:01:07.273724 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-purge-29331361-xd69z" Oct 08 00:01:07 crc kubenswrapper[4871]: I1008 00:01:07.273731 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04f7d8b72f555524da5e31144f107690d6a8949774b3d07f5940eb674d575dee" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.595132 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-khj7l"] Oct 08 00:01:28 crc kubenswrapper[4871]: E1008 00:01:28.596944 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dacc481-da68-45bf-b7a2-24f796328ee1" containerName="cinder-db-purge" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.596975 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dacc481-da68-45bf-b7a2-24f796328ee1" containerName="cinder-db-purge" Oct 08 00:01:28 crc kubenswrapper[4871]: E1008 00:01:28.597013 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35542928-206a-4b3f-bf2e-ba4ea4449cc3" containerName="keystone-cron" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.597032 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="35542928-206a-4b3f-bf2e-ba4ea4449cc3" containerName="keystone-cron" Oct 08 00:01:28 crc kubenswrapper[4871]: E1008 00:01:28.597108 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80141182-14e1-4535-a176-4c69d98eaaa5" containerName="glance-dbpurge" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.597127 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="80141182-14e1-4535-a176-4c69d98eaaa5" containerName="glance-dbpurge" Oct 08 00:01:28 crc kubenswrapper[4871]: E1008 00:01:28.597158 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401dc5ea-abe5-43e0-9d3f-095574675437" containerName="manila-db-purge" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.597176 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="401dc5ea-abe5-43e0-9d3f-095574675437" containerName="manila-db-purge" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.597742 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="80141182-14e1-4535-a176-4c69d98eaaa5" containerName="glance-dbpurge" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.597838 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dacc481-da68-45bf-b7a2-24f796328ee1" containerName="cinder-db-purge" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.597891 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="35542928-206a-4b3f-bf2e-ba4ea4449cc3" containerName="keystone-cron" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.597959 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="401dc5ea-abe5-43e0-9d3f-095574675437" containerName="manila-db-purge" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.601779 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.611622 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-khj7l"] Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.732057 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmqth\" (UniqueName: \"kubernetes.io/projected/495835c4-33dc-4564-8967-118a77d784fe-kube-api-access-tmqth\") pod \"redhat-marketplace-khj7l\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.732377 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-utilities\") pod \"redhat-marketplace-khj7l\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.732492 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-catalog-content\") pod \"redhat-marketplace-khj7l\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.834601 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmqth\" (UniqueName: \"kubernetes.io/projected/495835c4-33dc-4564-8967-118a77d784fe-kube-api-access-tmqth\") pod \"redhat-marketplace-khj7l\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.834757 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-utilities\") pod \"redhat-marketplace-khj7l\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.834902 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-catalog-content\") pod \"redhat-marketplace-khj7l\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.835541 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-catalog-content\") pod \"redhat-marketplace-khj7l\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.835755 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-utilities\") pod \"redhat-marketplace-khj7l\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.867033 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmqth\" (UniqueName: \"kubernetes.io/projected/495835c4-33dc-4564-8967-118a77d784fe-kube-api-access-tmqth\") pod \"redhat-marketplace-khj7l\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:28 crc kubenswrapper[4871]: I1008 00:01:28.955549 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:29 crc kubenswrapper[4871]: I1008 00:01:29.443511 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-khj7l"] Oct 08 00:01:29 crc kubenswrapper[4871]: I1008 00:01:29.535318 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khj7l" event={"ID":"495835c4-33dc-4564-8967-118a77d784fe","Type":"ContainerStarted","Data":"604ba157f393ed2d64d49df177ab8bdabfa94e2f5d15610c5edac2d99fb75821"} Oct 08 00:01:30 crc kubenswrapper[4871]: I1008 00:01:30.547548 4871 generic.go:334] "Generic (PLEG): container finished" podID="495835c4-33dc-4564-8967-118a77d784fe" containerID="9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85" exitCode=0 Oct 08 00:01:30 crc kubenswrapper[4871]: I1008 00:01:30.547706 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khj7l" event={"ID":"495835c4-33dc-4564-8967-118a77d784fe","Type":"ContainerDied","Data":"9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85"} Oct 08 00:01:31 crc kubenswrapper[4871]: I1008 00:01:31.562763 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khj7l" event={"ID":"495835c4-33dc-4564-8967-118a77d784fe","Type":"ContainerStarted","Data":"3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d"} Oct 08 00:01:32 crc kubenswrapper[4871]: I1008 00:01:32.576529 4871 generic.go:334] "Generic (PLEG): container finished" podID="495835c4-33dc-4564-8967-118a77d784fe" containerID="3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d" exitCode=0 Oct 08 00:01:32 crc kubenswrapper[4871]: I1008 00:01:32.576664 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khj7l" event={"ID":"495835c4-33dc-4564-8967-118a77d784fe","Type":"ContainerDied","Data":"3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d"} Oct 08 00:01:33 crc kubenswrapper[4871]: I1008 00:01:33.587454 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khj7l" event={"ID":"495835c4-33dc-4564-8967-118a77d784fe","Type":"ContainerStarted","Data":"ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79"} Oct 08 00:01:33 crc kubenswrapper[4871]: I1008 00:01:33.611613 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-khj7l" podStartSLOduration=3.194485619 podStartE2EDuration="5.611594398s" podCreationTimestamp="2025-10-08 00:01:28 +0000 UTC" firstStartedPulling="2025-10-08 00:01:30.54983477 +0000 UTC m=+6764.352532843" lastFinishedPulling="2025-10-08 00:01:32.966943519 +0000 UTC m=+6766.769641622" observedRunningTime="2025-10-08 00:01:33.606959206 +0000 UTC m=+6767.409657279" watchObservedRunningTime="2025-10-08 00:01:33.611594398 +0000 UTC m=+6767.414292471" Oct 08 00:01:38 crc kubenswrapper[4871]: I1008 00:01:38.956596 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:38 crc kubenswrapper[4871]: I1008 00:01:38.957255 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:39 crc kubenswrapper[4871]: I1008 00:01:39.026823 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:39 crc kubenswrapper[4871]: I1008 00:01:39.760947 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:39 crc kubenswrapper[4871]: I1008 00:01:39.896733 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-khj7l"] Oct 08 00:01:41 crc kubenswrapper[4871]: I1008 00:01:41.696109 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-khj7l" podUID="495835c4-33dc-4564-8967-118a77d784fe" containerName="registry-server" containerID="cri-o://ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79" gracePeriod=2 Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.239555 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.317191 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-utilities\") pod \"495835c4-33dc-4564-8967-118a77d784fe\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.317482 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmqth\" (UniqueName: \"kubernetes.io/projected/495835c4-33dc-4564-8967-118a77d784fe-kube-api-access-tmqth\") pod \"495835c4-33dc-4564-8967-118a77d784fe\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.317823 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-catalog-content\") pod \"495835c4-33dc-4564-8967-118a77d784fe\" (UID: \"495835c4-33dc-4564-8967-118a77d784fe\") " Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.318416 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-utilities" (OuterVolumeSpecName: "utilities") pod "495835c4-33dc-4564-8967-118a77d784fe" (UID: "495835c4-33dc-4564-8967-118a77d784fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.318633 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.331662 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/495835c4-33dc-4564-8967-118a77d784fe-kube-api-access-tmqth" (OuterVolumeSpecName: "kube-api-access-tmqth") pod "495835c4-33dc-4564-8967-118a77d784fe" (UID: "495835c4-33dc-4564-8967-118a77d784fe"). InnerVolumeSpecName "kube-api-access-tmqth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.332894 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "495835c4-33dc-4564-8967-118a77d784fe" (UID: "495835c4-33dc-4564-8967-118a77d784fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.421570 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmqth\" (UniqueName: \"kubernetes.io/projected/495835c4-33dc-4564-8967-118a77d784fe-kube-api-access-tmqth\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.421633 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495835c4-33dc-4564-8967-118a77d784fe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.711723 4871 generic.go:334] "Generic (PLEG): container finished" podID="495835c4-33dc-4564-8967-118a77d784fe" containerID="ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79" exitCode=0 Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.711781 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-khj7l" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.711811 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khj7l" event={"ID":"495835c4-33dc-4564-8967-118a77d784fe","Type":"ContainerDied","Data":"ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79"} Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.711984 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khj7l" event={"ID":"495835c4-33dc-4564-8967-118a77d784fe","Type":"ContainerDied","Data":"604ba157f393ed2d64d49df177ab8bdabfa94e2f5d15610c5edac2d99fb75821"} Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.712009 4871 scope.go:117] "RemoveContainer" containerID="ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.752651 4871 scope.go:117] "RemoveContainer" containerID="3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.759272 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-khj7l"] Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.775001 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-khj7l"] Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.799390 4871 scope.go:117] "RemoveContainer" containerID="9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.865765 4871 scope.go:117] "RemoveContainer" containerID="ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79" Oct 08 00:01:42 crc kubenswrapper[4871]: E1008 00:01:42.866464 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79\": container with ID starting with ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79 not found: ID does not exist" containerID="ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.866519 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79"} err="failed to get container status \"ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79\": rpc error: code = NotFound desc = could not find container \"ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79\": container with ID starting with ab3e8e7c263d22fb41fa980337b83eee8abc73e3f6427fc74c0b219b210a7e79 not found: ID does not exist" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.866553 4871 scope.go:117] "RemoveContainer" containerID="3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d" Oct 08 00:01:42 crc kubenswrapper[4871]: E1008 00:01:42.867200 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d\": container with ID starting with 3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d not found: ID does not exist" containerID="3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.867255 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d"} err="failed to get container status \"3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d\": rpc error: code = NotFound desc = could not find container \"3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d\": container with ID starting with 3fbf329cab4a0bd24d40e34d30a9d4fc800b5f5861cc54320833b349e9bb4d6d not found: ID does not exist" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.867291 4871 scope.go:117] "RemoveContainer" containerID="9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85" Oct 08 00:01:42 crc kubenswrapper[4871]: E1008 00:01:42.867655 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85\": container with ID starting with 9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85 not found: ID does not exist" containerID="9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85" Oct 08 00:01:42 crc kubenswrapper[4871]: I1008 00:01:42.867710 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85"} err="failed to get container status \"9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85\": rpc error: code = NotFound desc = could not find container \"9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85\": container with ID starting with 9821beaf41858bba7b26ba98af7563e9fd62d15d226b083b2757e17583fd2d85 not found: ID does not exist" Oct 08 00:01:43 crc kubenswrapper[4871]: I1008 00:01:43.000413 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="495835c4-33dc-4564-8967-118a77d784fe" path="/var/lib/kubelet/pods/495835c4-33dc-4564-8967-118a77d784fe/volumes" Oct 08 00:02:25 crc kubenswrapper[4871]: I1008 00:02:25.050078 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-pdkts"] Oct 08 00:02:25 crc kubenswrapper[4871]: I1008 00:02:25.065003 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-pdkts"] Oct 08 00:02:27 crc kubenswrapper[4871]: I1008 00:02:27.001426 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba09d7fe-5976-4a5b-8f7e-fbb0988e335f" path="/var/lib/kubelet/pods/ba09d7fe-5976-4a5b-8f7e-fbb0988e335f/volumes" Oct 08 00:02:35 crc kubenswrapper[4871]: I1008 00:02:35.049248 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-a7bc-account-create-f4rjs"] Oct 08 00:02:35 crc kubenswrapper[4871]: I1008 00:02:35.066197 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-a7bc-account-create-f4rjs"] Oct 08 00:02:35 crc kubenswrapper[4871]: I1008 00:02:35.512553 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:02:35 crc kubenswrapper[4871]: I1008 00:02:35.513043 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:02:37 crc kubenswrapper[4871]: I1008 00:02:37.002919 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c55d7c7-45dd-4e70-b1fe-cf549af69d58" path="/var/lib/kubelet/pods/5c55d7c7-45dd-4e70-b1fe-cf549af69d58/volumes" Oct 08 00:02:48 crc kubenswrapper[4871]: I1008 00:02:48.054500 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-f5qn6"] Oct 08 00:02:48 crc kubenswrapper[4871]: I1008 00:02:48.064129 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-f5qn6"] Oct 08 00:02:49 crc kubenswrapper[4871]: I1008 00:02:49.005474 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b72cc287-a1e9-4b19-a4b2-7ef096b6d324" path="/var/lib/kubelet/pods/b72cc287-a1e9-4b19-a4b2-7ef096b6d324/volumes" Oct 08 00:03:01 crc kubenswrapper[4871]: I1008 00:03:01.713648 4871 scope.go:117] "RemoveContainer" containerID="d9a8ad4187bab8ad6a9547b58dc52e9469af380516890ae99f8b40fc3a53f443" Oct 08 00:03:01 crc kubenswrapper[4871]: I1008 00:03:01.757430 4871 scope.go:117] "RemoveContainer" containerID="a0584b9a7c43e6906ad7b1643ab46010101ba18fa1118e85ed57db2ad65ae6c5" Oct 08 00:03:01 crc kubenswrapper[4871]: I1008 00:03:01.818372 4871 scope.go:117] "RemoveContainer" containerID="9a1263606d6d8da8935a1a8f8b6f784f7b769c8bd6f3b01b10d3b85f87e0b314" Oct 08 00:03:05 crc kubenswrapper[4871]: I1008 00:03:05.512179 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:03:05 crc kubenswrapper[4871]: I1008 00:03:05.512877 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:03:35 crc kubenswrapper[4871]: I1008 00:03:35.512723 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:03:35 crc kubenswrapper[4871]: I1008 00:03:35.513220 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:03:35 crc kubenswrapper[4871]: I1008 00:03:35.513256 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:03:35 crc kubenswrapper[4871]: I1008 00:03:35.514762 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"12ef4c6e132b32fbb6956c2b731cbfff15694db1157433b3657f2a0b245e4289"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:03:35 crc kubenswrapper[4871]: I1008 00:03:35.514825 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://12ef4c6e132b32fbb6956c2b731cbfff15694db1157433b3657f2a0b245e4289" gracePeriod=600 Oct 08 00:03:36 crc kubenswrapper[4871]: I1008 00:03:36.242579 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="12ef4c6e132b32fbb6956c2b731cbfff15694db1157433b3657f2a0b245e4289" exitCode=0 Oct 08 00:03:36 crc kubenswrapper[4871]: I1008 00:03:36.242690 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"12ef4c6e132b32fbb6956c2b731cbfff15694db1157433b3657f2a0b245e4289"} Oct 08 00:03:36 crc kubenswrapper[4871]: I1008 00:03:36.243467 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97"} Oct 08 00:03:36 crc kubenswrapper[4871]: I1008 00:03:36.243516 4871 scope.go:117] "RemoveContainer" containerID="eab732aa73fc56ff7b7a68f6ad890548875a880b5b091fe77afaa59cc16cd104" Oct 08 00:04:54 crc kubenswrapper[4871]: I1008 00:04:54.040192 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-gbqcd"] Oct 08 00:04:54 crc kubenswrapper[4871]: I1008 00:04:54.051268 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-gbqcd"] Oct 08 00:04:54 crc kubenswrapper[4871]: I1008 00:04:54.992818 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb8d68cc-caee-453d-8e02-9104fdb415c9" path="/var/lib/kubelet/pods/bb8d68cc-caee-453d-8e02-9104fdb415c9/volumes" Oct 08 00:05:01 crc kubenswrapper[4871]: I1008 00:05:01.961255 4871 scope.go:117] "RemoveContainer" containerID="1b4388d08b3a7187b520e69babee3c406b7c083955e0797d655ccecbfc3666c6" Oct 08 00:05:04 crc kubenswrapper[4871]: I1008 00:05:04.033101 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-e5a4-account-create-dshgj"] Oct 08 00:05:04 crc kubenswrapper[4871]: I1008 00:05:04.044063 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-e5a4-account-create-dshgj"] Oct 08 00:05:04 crc kubenswrapper[4871]: I1008 00:05:04.997131 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="079ddc46-d11f-4903-9bad-5d0753a722cd" path="/var/lib/kubelet/pods/079ddc46-d11f-4903-9bad-5d0753a722cd/volumes" Oct 08 00:05:17 crc kubenswrapper[4871]: I1008 00:05:17.045902 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-2m68z"] Oct 08 00:05:17 crc kubenswrapper[4871]: I1008 00:05:17.052962 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-2m68z"] Oct 08 00:05:18 crc kubenswrapper[4871]: I1008 00:05:18.996776 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e17c14-ed21-4d89-9299-da86a522d841" path="/var/lib/kubelet/pods/f7e17c14-ed21-4d89-9299-da86a522d841/volumes" Oct 08 00:05:35 crc kubenswrapper[4871]: I1008 00:05:35.512756 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:05:35 crc kubenswrapper[4871]: I1008 00:05:35.513319 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:05:37 crc kubenswrapper[4871]: I1008 00:05:37.044938 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-mgfvv"] Oct 08 00:05:37 crc kubenswrapper[4871]: I1008 00:05:37.056564 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-mgfvv"] Oct 08 00:05:38 crc kubenswrapper[4871]: I1008 00:05:38.997201 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24f2709b-5908-4273-a24b-53af7ff871da" path="/var/lib/kubelet/pods/24f2709b-5908-4273-a24b-53af7ff871da/volumes" Oct 08 00:05:40 crc kubenswrapper[4871]: I1008 00:05:40.898254 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9d4z5"] Oct 08 00:05:40 crc kubenswrapper[4871]: E1008 00:05:40.899753 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495835c4-33dc-4564-8967-118a77d784fe" containerName="registry-server" Oct 08 00:05:40 crc kubenswrapper[4871]: I1008 00:05:40.899772 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="495835c4-33dc-4564-8967-118a77d784fe" containerName="registry-server" Oct 08 00:05:40 crc kubenswrapper[4871]: E1008 00:05:40.899809 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495835c4-33dc-4564-8967-118a77d784fe" containerName="extract-utilities" Oct 08 00:05:40 crc kubenswrapper[4871]: I1008 00:05:40.899818 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="495835c4-33dc-4564-8967-118a77d784fe" containerName="extract-utilities" Oct 08 00:05:40 crc kubenswrapper[4871]: E1008 00:05:40.899837 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495835c4-33dc-4564-8967-118a77d784fe" containerName="extract-content" Oct 08 00:05:40 crc kubenswrapper[4871]: I1008 00:05:40.899845 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="495835c4-33dc-4564-8967-118a77d784fe" containerName="extract-content" Oct 08 00:05:40 crc kubenswrapper[4871]: I1008 00:05:40.900110 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="495835c4-33dc-4564-8967-118a77d784fe" containerName="registry-server" Oct 08 00:05:40 crc kubenswrapper[4871]: I1008 00:05:40.903752 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:40 crc kubenswrapper[4871]: I1008 00:05:40.958620 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9d4z5"] Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.055244 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-utilities\") pod \"certified-operators-9d4z5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.055302 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9wsk\" (UniqueName: \"kubernetes.io/projected/aa49491f-56fc-4f60-a663-65fea74756d5-kube-api-access-k9wsk\") pod \"certified-operators-9d4z5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.055365 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-catalog-content\") pod \"certified-operators-9d4z5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.158961 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-utilities\") pod \"certified-operators-9d4z5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.159027 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9wsk\" (UniqueName: \"kubernetes.io/projected/aa49491f-56fc-4f60-a663-65fea74756d5-kube-api-access-k9wsk\") pod \"certified-operators-9d4z5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.159110 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-catalog-content\") pod \"certified-operators-9d4z5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.160339 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-catalog-content\") pod \"certified-operators-9d4z5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.160603 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-utilities\") pod \"certified-operators-9d4z5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.184893 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9wsk\" (UniqueName: \"kubernetes.io/projected/aa49491f-56fc-4f60-a663-65fea74756d5-kube-api-access-k9wsk\") pod \"certified-operators-9d4z5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.243696 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:41 crc kubenswrapper[4871]: I1008 00:05:41.830043 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9d4z5"] Oct 08 00:05:42 crc kubenswrapper[4871]: I1008 00:05:42.698779 4871 generic.go:334] "Generic (PLEG): container finished" podID="aa49491f-56fc-4f60-a663-65fea74756d5" containerID="a3f482dbe1defb7fc89298230df38809d8d3c48aae330af7bc90ad782252140e" exitCode=0 Oct 08 00:05:42 crc kubenswrapper[4871]: I1008 00:05:42.698933 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9d4z5" event={"ID":"aa49491f-56fc-4f60-a663-65fea74756d5","Type":"ContainerDied","Data":"a3f482dbe1defb7fc89298230df38809d8d3c48aae330af7bc90ad782252140e"} Oct 08 00:05:42 crc kubenswrapper[4871]: I1008 00:05:42.699610 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9d4z5" event={"ID":"aa49491f-56fc-4f60-a663-65fea74756d5","Type":"ContainerStarted","Data":"d13ce97132075c02ff67c077345c0f4420258cabce86a9871a6e7b0e6764aec3"} Oct 08 00:05:42 crc kubenswrapper[4871]: I1008 00:05:42.705327 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:05:44 crc kubenswrapper[4871]: I1008 00:05:44.745549 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9d4z5" event={"ID":"aa49491f-56fc-4f60-a663-65fea74756d5","Type":"ContainerStarted","Data":"a3b6538031b78b95a83f0e4e3d19afdbdf82d626af47477928daebf64b250278"} Oct 08 00:05:45 crc kubenswrapper[4871]: I1008 00:05:45.760290 4871 generic.go:334] "Generic (PLEG): container finished" podID="aa49491f-56fc-4f60-a663-65fea74756d5" containerID="a3b6538031b78b95a83f0e4e3d19afdbdf82d626af47477928daebf64b250278" exitCode=0 Oct 08 00:05:45 crc kubenswrapper[4871]: I1008 00:05:45.760389 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9d4z5" event={"ID":"aa49491f-56fc-4f60-a663-65fea74756d5","Type":"ContainerDied","Data":"a3b6538031b78b95a83f0e4e3d19afdbdf82d626af47477928daebf64b250278"} Oct 08 00:05:46 crc kubenswrapper[4871]: I1008 00:05:46.773025 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9d4z5" event={"ID":"aa49491f-56fc-4f60-a663-65fea74756d5","Type":"ContainerStarted","Data":"26770fab8d75b353b38dbe41ea5847f1a6bb5b4d419cf4ddcb6af303b6c93299"} Oct 08 00:05:46 crc kubenswrapper[4871]: I1008 00:05:46.791622 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9d4z5" podStartSLOduration=3.205862617 podStartE2EDuration="6.791601698s" podCreationTimestamp="2025-10-08 00:05:40 +0000 UTC" firstStartedPulling="2025-10-08 00:05:42.70351161 +0000 UTC m=+7016.506209673" lastFinishedPulling="2025-10-08 00:05:46.289250681 +0000 UTC m=+7020.091948754" observedRunningTime="2025-10-08 00:05:46.790484828 +0000 UTC m=+7020.593182911" watchObservedRunningTime="2025-10-08 00:05:46.791601698 +0000 UTC m=+7020.594299771" Oct 08 00:05:47 crc kubenswrapper[4871]: I1008 00:05:47.028896 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-721c-account-create-nxz6g"] Oct 08 00:05:47 crc kubenswrapper[4871]: I1008 00:05:47.039476 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-721c-account-create-nxz6g"] Oct 08 00:05:48 crc kubenswrapper[4871]: I1008 00:05:48.996388 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54c505df-ac87-41ee-95cf-7e23638f1973" path="/var/lib/kubelet/pods/54c505df-ac87-41ee-95cf-7e23638f1973/volumes" Oct 08 00:05:51 crc kubenswrapper[4871]: I1008 00:05:51.244219 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:51 crc kubenswrapper[4871]: I1008 00:05:51.244671 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:51 crc kubenswrapper[4871]: I1008 00:05:51.333686 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:51 crc kubenswrapper[4871]: I1008 00:05:51.930297 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:51 crc kubenswrapper[4871]: I1008 00:05:51.992871 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9d4z5"] Oct 08 00:05:53 crc kubenswrapper[4871]: I1008 00:05:53.867764 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9d4z5" podUID="aa49491f-56fc-4f60-a663-65fea74756d5" containerName="registry-server" containerID="cri-o://26770fab8d75b353b38dbe41ea5847f1a6bb5b4d419cf4ddcb6af303b6c93299" gracePeriod=2 Oct 08 00:05:54 crc kubenswrapper[4871]: I1008 00:05:54.883141 4871 generic.go:334] "Generic (PLEG): container finished" podID="aa49491f-56fc-4f60-a663-65fea74756d5" containerID="26770fab8d75b353b38dbe41ea5847f1a6bb5b4d419cf4ddcb6af303b6c93299" exitCode=0 Oct 08 00:05:54 crc kubenswrapper[4871]: I1008 00:05:54.883225 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9d4z5" event={"ID":"aa49491f-56fc-4f60-a663-65fea74756d5","Type":"ContainerDied","Data":"26770fab8d75b353b38dbe41ea5847f1a6bb5b4d419cf4ddcb6af303b6c93299"} Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.383307 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.509038 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9wsk\" (UniqueName: \"kubernetes.io/projected/aa49491f-56fc-4f60-a663-65fea74756d5-kube-api-access-k9wsk\") pod \"aa49491f-56fc-4f60-a663-65fea74756d5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.509269 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-utilities\") pod \"aa49491f-56fc-4f60-a663-65fea74756d5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.509318 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-catalog-content\") pod \"aa49491f-56fc-4f60-a663-65fea74756d5\" (UID: \"aa49491f-56fc-4f60-a663-65fea74756d5\") " Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.510253 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-utilities" (OuterVolumeSpecName: "utilities") pod "aa49491f-56fc-4f60-a663-65fea74756d5" (UID: "aa49491f-56fc-4f60-a663-65fea74756d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.519078 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa49491f-56fc-4f60-a663-65fea74756d5-kube-api-access-k9wsk" (OuterVolumeSpecName: "kube-api-access-k9wsk") pod "aa49491f-56fc-4f60-a663-65fea74756d5" (UID: "aa49491f-56fc-4f60-a663-65fea74756d5"). InnerVolumeSpecName "kube-api-access-k9wsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.575705 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa49491f-56fc-4f60-a663-65fea74756d5" (UID: "aa49491f-56fc-4f60-a663-65fea74756d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.612182 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9wsk\" (UniqueName: \"kubernetes.io/projected/aa49491f-56fc-4f60-a663-65fea74756d5-kube-api-access-k9wsk\") on node \"crc\" DevicePath \"\"" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.612216 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.612225 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa49491f-56fc-4f60-a663-65fea74756d5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.893659 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9d4z5" event={"ID":"aa49491f-56fc-4f60-a663-65fea74756d5","Type":"ContainerDied","Data":"d13ce97132075c02ff67c077345c0f4420258cabce86a9871a6e7b0e6764aec3"} Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.893731 4871 scope.go:117] "RemoveContainer" containerID="26770fab8d75b353b38dbe41ea5847f1a6bb5b4d419cf4ddcb6af303b6c93299" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.895540 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9d4z5" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.925222 4871 scope.go:117] "RemoveContainer" containerID="a3b6538031b78b95a83f0e4e3d19afdbdf82d626af47477928daebf64b250278" Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.962920 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9d4z5"] Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.975336 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9d4z5"] Oct 08 00:05:55 crc kubenswrapper[4871]: I1008 00:05:55.975750 4871 scope.go:117] "RemoveContainer" containerID="a3f482dbe1defb7fc89298230df38809d8d3c48aae330af7bc90ad782252140e" Oct 08 00:05:56 crc kubenswrapper[4871]: I1008 00:05:56.993926 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa49491f-56fc-4f60-a663-65fea74756d5" path="/var/lib/kubelet/pods/aa49491f-56fc-4f60-a663-65fea74756d5/volumes" Oct 08 00:05:59 crc kubenswrapper[4871]: I1008 00:05:59.042891 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-8qtck"] Oct 08 00:05:59 crc kubenswrapper[4871]: I1008 00:05:59.053870 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-8qtck"] Oct 08 00:06:00 crc kubenswrapper[4871]: I1008 00:06:00.997069 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e18ea361-7369-4dca-827a-74d36b75bf7d" path="/var/lib/kubelet/pods/e18ea361-7369-4dca-827a-74d36b75bf7d/volumes" Oct 08 00:06:02 crc kubenswrapper[4871]: I1008 00:06:02.021574 4871 scope.go:117] "RemoveContainer" containerID="9dcbfcee077dff83a451ae4283810f0f19e9fbf6d90c826d630387d22cc7b359" Oct 08 00:06:02 crc kubenswrapper[4871]: I1008 00:06:02.047841 4871 scope.go:117] "RemoveContainer" containerID="042cf38dd89f92f56429b381c52fdebb20b5111ddf249243daede4db4265e2df" Oct 08 00:06:02 crc kubenswrapper[4871]: I1008 00:06:02.101346 4871 scope.go:117] "RemoveContainer" containerID="4d3692f0b1d79ab21fff7720b1014af96a533223d1e5841bb6c0564d6708f640" Oct 08 00:06:02 crc kubenswrapper[4871]: I1008 00:06:02.146632 4871 scope.go:117] "RemoveContainer" containerID="7f6eea7c4e0a143a41f6362051e5d54fde493ccd294027d317c82761c3b7a036" Oct 08 00:06:02 crc kubenswrapper[4871]: I1008 00:06:02.191186 4871 scope.go:117] "RemoveContainer" containerID="9051282a100a0f323719745bea42ef9a23c3813f748ca05f66e0f27884a1a792" Oct 08 00:06:02 crc kubenswrapper[4871]: I1008 00:06:02.214606 4871 scope.go:117] "RemoveContainer" containerID="79930ce111ddd8d4a05be42d44fc592b673016acdf7776e3ce1f6fd081079652" Oct 08 00:06:02 crc kubenswrapper[4871]: I1008 00:06:02.262166 4871 scope.go:117] "RemoveContainer" containerID="5e14d41415c1fb260632ce64ea841dfbe5acd9a67e2e210aa508333a06753076" Oct 08 00:06:02 crc kubenswrapper[4871]: I1008 00:06:02.335273 4871 scope.go:117] "RemoveContainer" containerID="d87db975bcc00bb201a13424f690fb19591106a0d107096d9360f72ea5b5cdb8" Oct 08 00:06:05 crc kubenswrapper[4871]: I1008 00:06:05.512867 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:06:05 crc kubenswrapper[4871]: I1008 00:06:05.513553 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:06:35 crc kubenswrapper[4871]: I1008 00:06:35.512173 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:06:35 crc kubenswrapper[4871]: I1008 00:06:35.512774 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:06:35 crc kubenswrapper[4871]: I1008 00:06:35.512860 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:06:35 crc kubenswrapper[4871]: I1008 00:06:35.513993 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:06:35 crc kubenswrapper[4871]: I1008 00:06:35.514073 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" gracePeriod=600 Oct 08 00:06:35 crc kubenswrapper[4871]: E1008 00:06:35.649195 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:06:36 crc kubenswrapper[4871]: I1008 00:06:36.383584 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" exitCode=0 Oct 08 00:06:36 crc kubenswrapper[4871]: I1008 00:06:36.383635 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97"} Oct 08 00:06:36 crc kubenswrapper[4871]: I1008 00:06:36.383678 4871 scope.go:117] "RemoveContainer" containerID="12ef4c6e132b32fbb6956c2b731cbfff15694db1157433b3657f2a0b245e4289" Oct 08 00:06:36 crc kubenswrapper[4871]: I1008 00:06:36.384597 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:06:36 crc kubenswrapper[4871]: E1008 00:06:36.385084 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:06:50 crc kubenswrapper[4871]: I1008 00:06:50.983143 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:06:50 crc kubenswrapper[4871]: E1008 00:06:50.984063 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:07:02 crc kubenswrapper[4871]: I1008 00:07:02.982880 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:07:02 crc kubenswrapper[4871]: E1008 00:07:02.983721 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:07:16 crc kubenswrapper[4871]: I1008 00:07:16.995223 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:07:16 crc kubenswrapper[4871]: E1008 00:07:16.996571 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:07:31 crc kubenswrapper[4871]: I1008 00:07:31.982724 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:07:31 crc kubenswrapper[4871]: E1008 00:07:31.983366 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:07:44 crc kubenswrapper[4871]: I1008 00:07:44.982849 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:07:44 crc kubenswrapper[4871]: E1008 00:07:44.985399 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:07:55 crc kubenswrapper[4871]: I1008 00:07:55.982916 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:07:55 crc kubenswrapper[4871]: E1008 00:07:55.983710 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:08:09 crc kubenswrapper[4871]: I1008 00:08:09.982731 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:08:09 crc kubenswrapper[4871]: E1008 00:08:09.983532 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:08:24 crc kubenswrapper[4871]: I1008 00:08:24.982904 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:08:25 crc kubenswrapper[4871]: E1008 00:08:25.001091 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.494892 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zldkg"] Oct 08 00:08:31 crc kubenswrapper[4871]: E1008 00:08:31.495838 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa49491f-56fc-4f60-a663-65fea74756d5" containerName="registry-server" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.495853 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa49491f-56fc-4f60-a663-65fea74756d5" containerName="registry-server" Oct 08 00:08:31 crc kubenswrapper[4871]: E1008 00:08:31.495877 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa49491f-56fc-4f60-a663-65fea74756d5" containerName="extract-utilities" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.495884 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa49491f-56fc-4f60-a663-65fea74756d5" containerName="extract-utilities" Oct 08 00:08:31 crc kubenswrapper[4871]: E1008 00:08:31.495906 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa49491f-56fc-4f60-a663-65fea74756d5" containerName="extract-content" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.495915 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa49491f-56fc-4f60-a663-65fea74756d5" containerName="extract-content" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.496207 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa49491f-56fc-4f60-a663-65fea74756d5" containerName="registry-server" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.497806 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.526270 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zldkg"] Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.674277 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-catalog-content\") pod \"community-operators-zldkg\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.674418 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghmqj\" (UniqueName: \"kubernetes.io/projected/ee82b15e-9132-452e-a049-9093d40818a9-kube-api-access-ghmqj\") pod \"community-operators-zldkg\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.674610 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-utilities\") pod \"community-operators-zldkg\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.777416 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-catalog-content\") pod \"community-operators-zldkg\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.777496 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghmqj\" (UniqueName: \"kubernetes.io/projected/ee82b15e-9132-452e-a049-9093d40818a9-kube-api-access-ghmqj\") pod \"community-operators-zldkg\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.777548 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-utilities\") pod \"community-operators-zldkg\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.778072 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-catalog-content\") pod \"community-operators-zldkg\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.778100 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-utilities\") pod \"community-operators-zldkg\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.812420 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghmqj\" (UniqueName: \"kubernetes.io/projected/ee82b15e-9132-452e-a049-9093d40818a9-kube-api-access-ghmqj\") pod \"community-operators-zldkg\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:31 crc kubenswrapper[4871]: I1008 00:08:31.820180 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:32 crc kubenswrapper[4871]: I1008 00:08:32.446563 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zldkg"] Oct 08 00:08:32 crc kubenswrapper[4871]: I1008 00:08:32.701255 4871 generic.go:334] "Generic (PLEG): container finished" podID="ee82b15e-9132-452e-a049-9093d40818a9" containerID="c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b" exitCode=0 Oct 08 00:08:32 crc kubenswrapper[4871]: I1008 00:08:32.701369 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zldkg" event={"ID":"ee82b15e-9132-452e-a049-9093d40818a9","Type":"ContainerDied","Data":"c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b"} Oct 08 00:08:32 crc kubenswrapper[4871]: I1008 00:08:32.701574 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zldkg" event={"ID":"ee82b15e-9132-452e-a049-9093d40818a9","Type":"ContainerStarted","Data":"a5277c2a0a12a596aa1fde46d670fc4b251209ac0b0148944466769274518cb7"} Oct 08 00:08:35 crc kubenswrapper[4871]: I1008 00:08:35.746670 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zldkg" event={"ID":"ee82b15e-9132-452e-a049-9093d40818a9","Type":"ContainerStarted","Data":"7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb"} Oct 08 00:08:36 crc kubenswrapper[4871]: I1008 00:08:36.989429 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:08:36 crc kubenswrapper[4871]: E1008 00:08:36.990236 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:08:37 crc kubenswrapper[4871]: I1008 00:08:37.771364 4871 generic.go:334] "Generic (PLEG): container finished" podID="ee82b15e-9132-452e-a049-9093d40818a9" containerID="7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb" exitCode=0 Oct 08 00:08:37 crc kubenswrapper[4871]: I1008 00:08:37.771423 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zldkg" event={"ID":"ee82b15e-9132-452e-a049-9093d40818a9","Type":"ContainerDied","Data":"7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb"} Oct 08 00:08:40 crc kubenswrapper[4871]: I1008 00:08:40.805504 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zldkg" event={"ID":"ee82b15e-9132-452e-a049-9093d40818a9","Type":"ContainerStarted","Data":"1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d"} Oct 08 00:08:40 crc kubenswrapper[4871]: I1008 00:08:40.828287 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zldkg" podStartSLOduration=2.759060621 podStartE2EDuration="9.828265173s" podCreationTimestamp="2025-10-08 00:08:31 +0000 UTC" firstStartedPulling="2025-10-08 00:08:32.703177533 +0000 UTC m=+7186.505875606" lastFinishedPulling="2025-10-08 00:08:39.772382055 +0000 UTC m=+7193.575080158" observedRunningTime="2025-10-08 00:08:40.82700372 +0000 UTC m=+7194.629701793" watchObservedRunningTime="2025-10-08 00:08:40.828265173 +0000 UTC m=+7194.630963256" Oct 08 00:08:41 crc kubenswrapper[4871]: I1008 00:08:41.821203 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:41 crc kubenswrapper[4871]: I1008 00:08:41.821556 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:42 crc kubenswrapper[4871]: I1008 00:08:42.867145 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zldkg" podUID="ee82b15e-9132-452e-a049-9093d40818a9" containerName="registry-server" probeResult="failure" output=< Oct 08 00:08:42 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 08 00:08:42 crc kubenswrapper[4871]: > Oct 08 00:08:51 crc kubenswrapper[4871]: I1008 00:08:51.899029 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:51 crc kubenswrapper[4871]: I1008 00:08:51.987538 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:08:51 crc kubenswrapper[4871]: E1008 00:08:51.987814 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:08:52 crc kubenswrapper[4871]: I1008 00:08:52.008897 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:52 crc kubenswrapper[4871]: I1008 00:08:52.148873 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zldkg"] Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:52.999838 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zldkg" podUID="ee82b15e-9132-452e-a049-9093d40818a9" containerName="registry-server" containerID="cri-o://1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d" gracePeriod=2 Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.590602 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.705957 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-utilities\") pod \"ee82b15e-9132-452e-a049-9093d40818a9\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.706113 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghmqj\" (UniqueName: \"kubernetes.io/projected/ee82b15e-9132-452e-a049-9093d40818a9-kube-api-access-ghmqj\") pod \"ee82b15e-9132-452e-a049-9093d40818a9\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.706157 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-catalog-content\") pod \"ee82b15e-9132-452e-a049-9093d40818a9\" (UID: \"ee82b15e-9132-452e-a049-9093d40818a9\") " Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.707286 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-utilities" (OuterVolumeSpecName: "utilities") pod "ee82b15e-9132-452e-a049-9093d40818a9" (UID: "ee82b15e-9132-452e-a049-9093d40818a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.716127 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee82b15e-9132-452e-a049-9093d40818a9-kube-api-access-ghmqj" (OuterVolumeSpecName: "kube-api-access-ghmqj") pod "ee82b15e-9132-452e-a049-9093d40818a9" (UID: "ee82b15e-9132-452e-a049-9093d40818a9"). InnerVolumeSpecName "kube-api-access-ghmqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.775235 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee82b15e-9132-452e-a049-9093d40818a9" (UID: "ee82b15e-9132-452e-a049-9093d40818a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.809402 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.809473 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghmqj\" (UniqueName: \"kubernetes.io/projected/ee82b15e-9132-452e-a049-9093d40818a9-kube-api-access-ghmqj\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:53 crc kubenswrapper[4871]: I1008 00:08:53.809496 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee82b15e-9132-452e-a049-9093d40818a9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.014868 4871 generic.go:334] "Generic (PLEG): container finished" podID="ee82b15e-9132-452e-a049-9093d40818a9" containerID="1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d" exitCode=0 Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.014943 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zldkg" event={"ID":"ee82b15e-9132-452e-a049-9093d40818a9","Type":"ContainerDied","Data":"1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d"} Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.014971 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zldkg" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.015013 4871 scope.go:117] "RemoveContainer" containerID="1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.015000 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zldkg" event={"ID":"ee82b15e-9132-452e-a049-9093d40818a9","Type":"ContainerDied","Data":"a5277c2a0a12a596aa1fde46d670fc4b251209ac0b0148944466769274518cb7"} Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.045663 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zldkg"] Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.050348 4871 scope.go:117] "RemoveContainer" containerID="7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.053372 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zldkg"] Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.079502 4871 scope.go:117] "RemoveContainer" containerID="c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.141052 4871 scope.go:117] "RemoveContainer" containerID="1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d" Oct 08 00:08:54 crc kubenswrapper[4871]: E1008 00:08:54.141617 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d\": container with ID starting with 1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d not found: ID does not exist" containerID="1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.141659 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d"} err="failed to get container status \"1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d\": rpc error: code = NotFound desc = could not find container \"1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d\": container with ID starting with 1d3a6897945b7ae24440c54202b347575f2742ad5fc3741a1fd14a6f79c1c68d not found: ID does not exist" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.141685 4871 scope.go:117] "RemoveContainer" containerID="7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb" Oct 08 00:08:54 crc kubenswrapper[4871]: E1008 00:08:54.142247 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb\": container with ID starting with 7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb not found: ID does not exist" containerID="7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.142271 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb"} err="failed to get container status \"7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb\": rpc error: code = NotFound desc = could not find container \"7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb\": container with ID starting with 7ad1c5e3ac61f372f86b50c9ac9e5dd1cd6000e9988e2e6cff9673b09d1c73fb not found: ID does not exist" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.142285 4871 scope.go:117] "RemoveContainer" containerID="c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b" Oct 08 00:08:54 crc kubenswrapper[4871]: E1008 00:08:54.142621 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b\": container with ID starting with c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b not found: ID does not exist" containerID="c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b" Oct 08 00:08:54 crc kubenswrapper[4871]: I1008 00:08:54.142692 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b"} err="failed to get container status \"c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b\": rpc error: code = NotFound desc = could not find container \"c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b\": container with ID starting with c60b2989ce7ef6a11e1a70e5ecf3d31600103f6de87d6e9707eb3fa7e2c7e96b not found: ID does not exist" Oct 08 00:08:55 crc kubenswrapper[4871]: I1008 00:08:55.000148 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee82b15e-9132-452e-a049-9093d40818a9" path="/var/lib/kubelet/pods/ee82b15e-9132-452e-a049-9093d40818a9/volumes" Oct 08 00:08:55 crc kubenswrapper[4871]: I1008 00:08:55.031451 4871 generic.go:334] "Generic (PLEG): container finished" podID="8a39a36c-8c17-4049-99a0-3ef681c082b3" containerID="bfe0d5210a49c7f33d9a443c7a9de84809e4a127626addc0ad4d4885795a1a0c" exitCode=0 Oct 08 00:08:55 crc kubenswrapper[4871]: I1008 00:08:55.031565 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" event={"ID":"8a39a36c-8c17-4049-99a0-3ef681c082b3","Type":"ContainerDied","Data":"bfe0d5210a49c7f33d9a443c7a9de84809e4a127626addc0ad4d4885795a1a0c"} Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.609880 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.790438 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-tripleo-cleanup-combined-ca-bundle\") pod \"8a39a36c-8c17-4049-99a0-3ef681c082b3\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.790517 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-inventory\") pod \"8a39a36c-8c17-4049-99a0-3ef681c082b3\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.790623 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ssh-key\") pod \"8a39a36c-8c17-4049-99a0-3ef681c082b3\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.790657 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ceph\") pod \"8a39a36c-8c17-4049-99a0-3ef681c082b3\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.790682 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8b58\" (UniqueName: \"kubernetes.io/projected/8a39a36c-8c17-4049-99a0-3ef681c082b3-kube-api-access-b8b58\") pod \"8a39a36c-8c17-4049-99a0-3ef681c082b3\" (UID: \"8a39a36c-8c17-4049-99a0-3ef681c082b3\") " Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.797415 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ceph" (OuterVolumeSpecName: "ceph") pod "8a39a36c-8c17-4049-99a0-3ef681c082b3" (UID: "8a39a36c-8c17-4049-99a0-3ef681c082b3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.797483 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "8a39a36c-8c17-4049-99a0-3ef681c082b3" (UID: "8a39a36c-8c17-4049-99a0-3ef681c082b3"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.798437 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a39a36c-8c17-4049-99a0-3ef681c082b3-kube-api-access-b8b58" (OuterVolumeSpecName: "kube-api-access-b8b58") pod "8a39a36c-8c17-4049-99a0-3ef681c082b3" (UID: "8a39a36c-8c17-4049-99a0-3ef681c082b3"). InnerVolumeSpecName "kube-api-access-b8b58". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.844306 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a39a36c-8c17-4049-99a0-3ef681c082b3" (UID: "8a39a36c-8c17-4049-99a0-3ef681c082b3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.844649 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-inventory" (OuterVolumeSpecName: "inventory") pod "8a39a36c-8c17-4049-99a0-3ef681c082b3" (UID: "8a39a36c-8c17-4049-99a0-3ef681c082b3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.894296 4871 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.894339 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.894352 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.894364 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a39a36c-8c17-4049-99a0-3ef681c082b3-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:56 crc kubenswrapper[4871]: I1008 00:08:56.894378 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8b58\" (UniqueName: \"kubernetes.io/projected/8a39a36c-8c17-4049-99a0-3ef681c082b3-kube-api-access-b8b58\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:57 crc kubenswrapper[4871]: I1008 00:08:57.058834 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" event={"ID":"8a39a36c-8c17-4049-99a0-3ef681c082b3","Type":"ContainerDied","Data":"189c803ff3ad571a4ff3600411123249b1820bc1be19c0cf56a6684aa0eefb35"} Oct 08 00:08:57 crc kubenswrapper[4871]: I1008 00:08:57.059091 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="189c803ff3ad571a4ff3600411123249b1820bc1be19c0cf56a6684aa0eefb35" Oct 08 00:08:57 crc kubenswrapper[4871]: I1008 00:08:57.058908 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.963507 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-xxtff"] Oct 08 00:09:01 crc kubenswrapper[4871]: E1008 00:09:01.965476 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee82b15e-9132-452e-a049-9093d40818a9" containerName="extract-utilities" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.965502 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee82b15e-9132-452e-a049-9093d40818a9" containerName="extract-utilities" Oct 08 00:09:01 crc kubenswrapper[4871]: E1008 00:09:01.965523 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a39a36c-8c17-4049-99a0-3ef681c082b3" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.965533 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a39a36c-8c17-4049-99a0-3ef681c082b3" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 08 00:09:01 crc kubenswrapper[4871]: E1008 00:09:01.965563 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee82b15e-9132-452e-a049-9093d40818a9" containerName="extract-content" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.965571 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee82b15e-9132-452e-a049-9093d40818a9" containerName="extract-content" Oct 08 00:09:01 crc kubenswrapper[4871]: E1008 00:09:01.965619 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee82b15e-9132-452e-a049-9093d40818a9" containerName="registry-server" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.965627 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee82b15e-9132-452e-a049-9093d40818a9" containerName="registry-server" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.966182 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee82b15e-9132-452e-a049-9093d40818a9" containerName="registry-server" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.966215 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a39a36c-8c17-4049-99a0-3ef681c082b3" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.967281 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.970013 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.970548 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.970627 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.970632 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:09:01 crc kubenswrapper[4871]: I1008 00:09:01.992261 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-xxtff"] Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.024026 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7lh6\" (UniqueName: \"kubernetes.io/projected/c5eb167d-5c30-4354-888c-25c58fdfd2e9-kube-api-access-c7lh6\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.024083 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ceph\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.024111 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-inventory\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.024148 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.024206 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.125964 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7lh6\" (UniqueName: \"kubernetes.io/projected/c5eb167d-5c30-4354-888c-25c58fdfd2e9-kube-api-access-c7lh6\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.126019 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ceph\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.126052 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-inventory\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.126088 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.126151 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.132742 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-inventory\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.134703 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.135573 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ceph\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.140624 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.144722 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7lh6\" (UniqueName: \"kubernetes.io/projected/c5eb167d-5c30-4354-888c-25c58fdfd2e9-kube-api-access-c7lh6\") pod \"bootstrap-openstack-openstack-cell1-xxtff\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.297297 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:09:02 crc kubenswrapper[4871]: I1008 00:09:02.890280 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-xxtff"] Oct 08 00:09:02 crc kubenswrapper[4871]: W1008 00:09:02.903551 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5eb167d_5c30_4354_888c_25c58fdfd2e9.slice/crio-bf1d63c37ac1209b2a0243217be01a4f06cab5db9d41d1c94f452b3bd66db860 WatchSource:0}: Error finding container bf1d63c37ac1209b2a0243217be01a4f06cab5db9d41d1c94f452b3bd66db860: Status 404 returned error can't find the container with id bf1d63c37ac1209b2a0243217be01a4f06cab5db9d41d1c94f452b3bd66db860 Oct 08 00:09:03 crc kubenswrapper[4871]: I1008 00:09:03.123909 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" event={"ID":"c5eb167d-5c30-4354-888c-25c58fdfd2e9","Type":"ContainerStarted","Data":"bf1d63c37ac1209b2a0243217be01a4f06cab5db9d41d1c94f452b3bd66db860"} Oct 08 00:09:03 crc kubenswrapper[4871]: I1008 00:09:03.982395 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:09:03 crc kubenswrapper[4871]: E1008 00:09:03.983035 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:09:05 crc kubenswrapper[4871]: I1008 00:09:05.153962 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" event={"ID":"c5eb167d-5c30-4354-888c-25c58fdfd2e9","Type":"ContainerStarted","Data":"7e76ebfa725b7599ae08104f12cdaa75a5630fdfdfea6544a235737ec6678d5c"} Oct 08 00:09:05 crc kubenswrapper[4871]: I1008 00:09:05.185763 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" podStartSLOduration=2.5046341869999997 podStartE2EDuration="4.185743397s" podCreationTimestamp="2025-10-08 00:09:01 +0000 UTC" firstStartedPulling="2025-10-08 00:09:02.935850376 +0000 UTC m=+7216.738548469" lastFinishedPulling="2025-10-08 00:09:04.616959566 +0000 UTC m=+7218.419657679" observedRunningTime="2025-10-08 00:09:05.179266037 +0000 UTC m=+7218.981964140" watchObservedRunningTime="2025-10-08 00:09:05.185743397 +0000 UTC m=+7218.988441460" Oct 08 00:09:14 crc kubenswrapper[4871]: I1008 00:09:14.982283 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:09:14 crc kubenswrapper[4871]: E1008 00:09:14.983016 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:09:28 crc kubenswrapper[4871]: I1008 00:09:28.982611 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:09:28 crc kubenswrapper[4871]: E1008 00:09:28.983319 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:09:41 crc kubenswrapper[4871]: I1008 00:09:41.983103 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:09:41 crc kubenswrapper[4871]: E1008 00:09:41.983928 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.440119 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hjwwp"] Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.445114 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.453723 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hjwwp"] Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.531952 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-catalog-content\") pod \"redhat-operators-hjwwp\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.531997 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nblz9\" (UniqueName: \"kubernetes.io/projected/a0186640-5732-495f-a98a-74deb0eb56c1-kube-api-access-nblz9\") pod \"redhat-operators-hjwwp\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.532254 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-utilities\") pod \"redhat-operators-hjwwp\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.633491 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-utilities\") pod \"redhat-operators-hjwwp\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.633653 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-catalog-content\") pod \"redhat-operators-hjwwp\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.633684 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nblz9\" (UniqueName: \"kubernetes.io/projected/a0186640-5732-495f-a98a-74deb0eb56c1-kube-api-access-nblz9\") pod \"redhat-operators-hjwwp\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.634202 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-utilities\") pod \"redhat-operators-hjwwp\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.634246 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-catalog-content\") pod \"redhat-operators-hjwwp\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.654267 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nblz9\" (UniqueName: \"kubernetes.io/projected/a0186640-5732-495f-a98a-74deb0eb56c1-kube-api-access-nblz9\") pod \"redhat-operators-hjwwp\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.781508 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:09:53 crc kubenswrapper[4871]: I1008 00:09:53.982334 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:09:53 crc kubenswrapper[4871]: E1008 00:09:53.982668 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:09:54 crc kubenswrapper[4871]: I1008 00:09:54.344951 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hjwwp"] Oct 08 00:09:54 crc kubenswrapper[4871]: I1008 00:09:54.756683 4871 generic.go:334] "Generic (PLEG): container finished" podID="a0186640-5732-495f-a98a-74deb0eb56c1" containerID="31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6" exitCode=0 Oct 08 00:09:54 crc kubenswrapper[4871]: I1008 00:09:54.756785 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjwwp" event={"ID":"a0186640-5732-495f-a98a-74deb0eb56c1","Type":"ContainerDied","Data":"31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6"} Oct 08 00:09:54 crc kubenswrapper[4871]: I1008 00:09:54.757142 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjwwp" event={"ID":"a0186640-5732-495f-a98a-74deb0eb56c1","Type":"ContainerStarted","Data":"45bb0a61f55556173a013ce24b3cb61418ca6853b074e1bb76d070067ea609b1"} Oct 08 00:09:56 crc kubenswrapper[4871]: I1008 00:09:56.803920 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjwwp" event={"ID":"a0186640-5732-495f-a98a-74deb0eb56c1","Type":"ContainerStarted","Data":"4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02"} Oct 08 00:09:59 crc kubenswrapper[4871]: I1008 00:09:59.840887 4871 generic.go:334] "Generic (PLEG): container finished" podID="a0186640-5732-495f-a98a-74deb0eb56c1" containerID="4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02" exitCode=0 Oct 08 00:09:59 crc kubenswrapper[4871]: I1008 00:09:59.840977 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjwwp" event={"ID":"a0186640-5732-495f-a98a-74deb0eb56c1","Type":"ContainerDied","Data":"4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02"} Oct 08 00:10:00 crc kubenswrapper[4871]: I1008 00:10:00.873263 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjwwp" event={"ID":"a0186640-5732-495f-a98a-74deb0eb56c1","Type":"ContainerStarted","Data":"d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f"} Oct 08 00:10:00 crc kubenswrapper[4871]: I1008 00:10:00.928994 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hjwwp" podStartSLOduration=2.200769439 podStartE2EDuration="7.928971583s" podCreationTimestamp="2025-10-08 00:09:53 +0000 UTC" firstStartedPulling="2025-10-08 00:09:54.758343063 +0000 UTC m=+7268.561041146" lastFinishedPulling="2025-10-08 00:10:00.486545217 +0000 UTC m=+7274.289243290" observedRunningTime="2025-10-08 00:10:00.917506372 +0000 UTC m=+7274.720204455" watchObservedRunningTime="2025-10-08 00:10:00.928971583 +0000 UTC m=+7274.731669656" Oct 08 00:10:03 crc kubenswrapper[4871]: I1008 00:10:03.782411 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:10:03 crc kubenswrapper[4871]: I1008 00:10:03.782711 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:10:04 crc kubenswrapper[4871]: I1008 00:10:04.837895 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hjwwp" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" containerName="registry-server" probeResult="failure" output=< Oct 08 00:10:04 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 08 00:10:04 crc kubenswrapper[4871]: > Oct 08 00:10:05 crc kubenswrapper[4871]: I1008 00:10:05.982624 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:10:05 crc kubenswrapper[4871]: E1008 00:10:05.983037 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:10:13 crc kubenswrapper[4871]: I1008 00:10:13.858991 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:10:13 crc kubenswrapper[4871]: I1008 00:10:13.928078 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:10:14 crc kubenswrapper[4871]: I1008 00:10:14.107602 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hjwwp"] Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.049114 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hjwwp" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" containerName="registry-server" containerID="cri-o://d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f" gracePeriod=2 Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.616282 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.704451 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-catalog-content\") pod \"a0186640-5732-495f-a98a-74deb0eb56c1\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.705048 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nblz9\" (UniqueName: \"kubernetes.io/projected/a0186640-5732-495f-a98a-74deb0eb56c1-kube-api-access-nblz9\") pod \"a0186640-5732-495f-a98a-74deb0eb56c1\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.705081 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-utilities\") pod \"a0186640-5732-495f-a98a-74deb0eb56c1\" (UID: \"a0186640-5732-495f-a98a-74deb0eb56c1\") " Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.706615 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-utilities" (OuterVolumeSpecName: "utilities") pod "a0186640-5732-495f-a98a-74deb0eb56c1" (UID: "a0186640-5732-495f-a98a-74deb0eb56c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.713050 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0186640-5732-495f-a98a-74deb0eb56c1-kube-api-access-nblz9" (OuterVolumeSpecName: "kube-api-access-nblz9") pod "a0186640-5732-495f-a98a-74deb0eb56c1" (UID: "a0186640-5732-495f-a98a-74deb0eb56c1"). InnerVolumeSpecName "kube-api-access-nblz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.792362 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0186640-5732-495f-a98a-74deb0eb56c1" (UID: "a0186640-5732-495f-a98a-74deb0eb56c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.808273 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.808303 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nblz9\" (UniqueName: \"kubernetes.io/projected/a0186640-5732-495f-a98a-74deb0eb56c1-kube-api-access-nblz9\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:15 crc kubenswrapper[4871]: I1008 00:10:15.808313 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0186640-5732-495f-a98a-74deb0eb56c1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.060393 4871 generic.go:334] "Generic (PLEG): container finished" podID="a0186640-5732-495f-a98a-74deb0eb56c1" containerID="d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f" exitCode=0 Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.060449 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjwwp" event={"ID":"a0186640-5732-495f-a98a-74deb0eb56c1","Type":"ContainerDied","Data":"d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f"} Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.060488 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjwwp" event={"ID":"a0186640-5732-495f-a98a-74deb0eb56c1","Type":"ContainerDied","Data":"45bb0a61f55556173a013ce24b3cb61418ca6853b074e1bb76d070067ea609b1"} Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.060512 4871 scope.go:117] "RemoveContainer" containerID="d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.060561 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hjwwp" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.087202 4871 scope.go:117] "RemoveContainer" containerID="4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.130997 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hjwwp"] Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.137319 4871 scope.go:117] "RemoveContainer" containerID="31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.148336 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hjwwp"] Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.177369 4871 scope.go:117] "RemoveContainer" containerID="d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f" Oct 08 00:10:16 crc kubenswrapper[4871]: E1008 00:10:16.178140 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f\": container with ID starting with d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f not found: ID does not exist" containerID="d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.178188 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f"} err="failed to get container status \"d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f\": rpc error: code = NotFound desc = could not find container \"d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f\": container with ID starting with d339eca3ff358a54a72a4d1de3af0ad439b3a90e2a2124e75c34c0332a2df82f not found: ID does not exist" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.178220 4871 scope.go:117] "RemoveContainer" containerID="4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02" Oct 08 00:10:16 crc kubenswrapper[4871]: E1008 00:10:16.178732 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02\": container with ID starting with 4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02 not found: ID does not exist" containerID="4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.178998 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02"} err="failed to get container status \"4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02\": rpc error: code = NotFound desc = could not find container \"4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02\": container with ID starting with 4fcf74d2ce7078356a3f3a8134cbb5ef5578b1abaccf7705cb4d71c9aa6c8c02 not found: ID does not exist" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.179118 4871 scope.go:117] "RemoveContainer" containerID="31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6" Oct 08 00:10:16 crc kubenswrapper[4871]: E1008 00:10:16.179751 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6\": container with ID starting with 31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6 not found: ID does not exist" containerID="31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.179819 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6"} err="failed to get container status \"31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6\": rpc error: code = NotFound desc = could not find container \"31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6\": container with ID starting with 31841b53baff54eef92d41e6d9a2e37a70ebae09472a7d27f5ffa57783f76ac6 not found: ID does not exist" Oct 08 00:10:16 crc kubenswrapper[4871]: I1008 00:10:16.999352 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" path="/var/lib/kubelet/pods/a0186640-5732-495f-a98a-74deb0eb56c1/volumes" Oct 08 00:10:20 crc kubenswrapper[4871]: I1008 00:10:20.982905 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:10:20 crc kubenswrapper[4871]: E1008 00:10:20.983469 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:10:33 crc kubenswrapper[4871]: I1008 00:10:33.983458 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:10:33 crc kubenswrapper[4871]: E1008 00:10:33.984633 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:10:45 crc kubenswrapper[4871]: I1008 00:10:45.982553 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:10:45 crc kubenswrapper[4871]: E1008 00:10:45.983602 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:10:59 crc kubenswrapper[4871]: I1008 00:10:59.983331 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:10:59 crc kubenswrapper[4871]: E1008 00:10:59.984204 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:11:12 crc kubenswrapper[4871]: I1008 00:11:12.982547 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:11:12 crc kubenswrapper[4871]: E1008 00:11:12.984194 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:11:25 crc kubenswrapper[4871]: I1008 00:11:25.982966 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:11:25 crc kubenswrapper[4871]: E1008 00:11:25.984944 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:11:39 crc kubenswrapper[4871]: I1008 00:11:39.982492 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:11:41 crc kubenswrapper[4871]: I1008 00:11:41.044968 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"032447e2c741f3b9f493330ad578f9e126416d12ba3cc2e1856ed60479a27106"} Oct 08 00:12:25 crc kubenswrapper[4871]: I1008 00:12:25.552226 4871 generic.go:334] "Generic (PLEG): container finished" podID="c5eb167d-5c30-4354-888c-25c58fdfd2e9" containerID="7e76ebfa725b7599ae08104f12cdaa75a5630fdfdfea6544a235737ec6678d5c" exitCode=0 Oct 08 00:12:25 crc kubenswrapper[4871]: I1008 00:12:25.552294 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" event={"ID":"c5eb167d-5c30-4354-888c-25c58fdfd2e9","Type":"ContainerDied","Data":"7e76ebfa725b7599ae08104f12cdaa75a5630fdfdfea6544a235737ec6678d5c"} Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.871152 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mf2vl"] Oct 08 00:12:26 crc kubenswrapper[4871]: E1008 00:12:26.872129 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" containerName="registry-server" Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.872146 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" containerName="registry-server" Oct 08 00:12:26 crc kubenswrapper[4871]: E1008 00:12:26.872168 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" containerName="extract-content" Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.872176 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" containerName="extract-content" Oct 08 00:12:26 crc kubenswrapper[4871]: E1008 00:12:26.872200 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" containerName="extract-utilities" Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.872210 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" containerName="extract-utilities" Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.872640 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0186640-5732-495f-a98a-74deb0eb56c1" containerName="registry-server" Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.875256 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.897467 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mf2vl"] Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.915580 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-catalog-content\") pod \"redhat-marketplace-mf2vl\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.915678 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-utilities\") pod \"redhat-marketplace-mf2vl\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:26 crc kubenswrapper[4871]: I1008 00:12:26.915787 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjv5c\" (UniqueName: \"kubernetes.io/projected/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-kube-api-access-pjv5c\") pod \"redhat-marketplace-mf2vl\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.019462 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjv5c\" (UniqueName: \"kubernetes.io/projected/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-kube-api-access-pjv5c\") pod \"redhat-marketplace-mf2vl\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.019587 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-catalog-content\") pod \"redhat-marketplace-mf2vl\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.019628 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-utilities\") pod \"redhat-marketplace-mf2vl\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.020243 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-utilities\") pod \"redhat-marketplace-mf2vl\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.020708 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-catalog-content\") pod \"redhat-marketplace-mf2vl\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.052372 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjv5c\" (UniqueName: \"kubernetes.io/projected/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-kube-api-access-pjv5c\") pod \"redhat-marketplace-mf2vl\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.118458 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.223037 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-bootstrap-combined-ca-bundle\") pod \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.223120 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-inventory\") pod \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.223263 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ceph\") pod \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.223281 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7lh6\" (UniqueName: \"kubernetes.io/projected/c5eb167d-5c30-4354-888c-25c58fdfd2e9-kube-api-access-c7lh6\") pod \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.223312 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ssh-key\") pod \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\" (UID: \"c5eb167d-5c30-4354-888c-25c58fdfd2e9\") " Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.225023 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.227980 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c5eb167d-5c30-4354-888c-25c58fdfd2e9" (UID: "c5eb167d-5c30-4354-888c-25c58fdfd2e9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.231974 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5eb167d-5c30-4354-888c-25c58fdfd2e9-kube-api-access-c7lh6" (OuterVolumeSpecName: "kube-api-access-c7lh6") pod "c5eb167d-5c30-4354-888c-25c58fdfd2e9" (UID: "c5eb167d-5c30-4354-888c-25c58fdfd2e9"). InnerVolumeSpecName "kube-api-access-c7lh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.231992 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ceph" (OuterVolumeSpecName: "ceph") pod "c5eb167d-5c30-4354-888c-25c58fdfd2e9" (UID: "c5eb167d-5c30-4354-888c-25c58fdfd2e9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.264571 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-inventory" (OuterVolumeSpecName: "inventory") pod "c5eb167d-5c30-4354-888c-25c58fdfd2e9" (UID: "c5eb167d-5c30-4354-888c-25c58fdfd2e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.293519 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c5eb167d-5c30-4354-888c-25c58fdfd2e9" (UID: "c5eb167d-5c30-4354-888c-25c58fdfd2e9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.336840 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.336874 4871 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.336887 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.336896 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5eb167d-5c30-4354-888c-25c58fdfd2e9-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.336904 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7lh6\" (UniqueName: \"kubernetes.io/projected/c5eb167d-5c30-4354-888c-25c58fdfd2e9-kube-api-access-c7lh6\") on node \"crc\" DevicePath \"\"" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.572403 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" event={"ID":"c5eb167d-5c30-4354-888c-25c58fdfd2e9","Type":"ContainerDied","Data":"bf1d63c37ac1209b2a0243217be01a4f06cab5db9d41d1c94f452b3bd66db860"} Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.572447 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf1d63c37ac1209b2a0243217be01a4f06cab5db9d41d1c94f452b3bd66db860" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.572454 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-xxtff" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.732772 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mf2vl"] Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.743120 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-td4fb"] Oct 08 00:12:27 crc kubenswrapper[4871]: E1008 00:12:27.743878 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5eb167d-5c30-4354-888c-25c58fdfd2e9" containerName="bootstrap-openstack-openstack-cell1" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.743898 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5eb167d-5c30-4354-888c-25c58fdfd2e9" containerName="bootstrap-openstack-openstack-cell1" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.744140 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5eb167d-5c30-4354-888c-25c58fdfd2e9" containerName="bootstrap-openstack-openstack-cell1" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.745063 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.749376 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.749399 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.749560 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.749631 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.778220 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-td4fb"] Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.847271 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ceph\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.847442 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-inventory\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.847495 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdwqr\" (UniqueName: \"kubernetes.io/projected/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-kube-api-access-vdwqr\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.847537 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ssh-key\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.949033 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-inventory\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.949092 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdwqr\" (UniqueName: \"kubernetes.io/projected/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-kube-api-access-vdwqr\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.949157 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ssh-key\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.949258 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ceph\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.955043 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ceph\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.955583 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-inventory\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.955723 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ssh-key\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:27 crc kubenswrapper[4871]: I1008 00:12:27.968011 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdwqr\" (UniqueName: \"kubernetes.io/projected/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-kube-api-access-vdwqr\") pod \"download-cache-openstack-openstack-cell1-td4fb\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:28 crc kubenswrapper[4871]: I1008 00:12:28.067371 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:12:28 crc kubenswrapper[4871]: I1008 00:12:28.586198 4871 generic.go:334] "Generic (PLEG): container finished" podID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerID="d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad" exitCode=0 Oct 08 00:12:28 crc kubenswrapper[4871]: I1008 00:12:28.586284 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mf2vl" event={"ID":"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2","Type":"ContainerDied","Data":"d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad"} Oct 08 00:12:28 crc kubenswrapper[4871]: I1008 00:12:28.586621 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mf2vl" event={"ID":"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2","Type":"ContainerStarted","Data":"61a9a8e74aa2873a848d271462f30fe487da5d21194e15e9daa9296bb172cdd1"} Oct 08 00:12:28 crc kubenswrapper[4871]: I1008 00:12:28.589291 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:12:28 crc kubenswrapper[4871]: I1008 00:12:28.640288 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-td4fb"] Oct 08 00:12:29 crc kubenswrapper[4871]: I1008 00:12:29.595503 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-td4fb" event={"ID":"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793","Type":"ContainerStarted","Data":"325f401a9299f9ef23fe7519b4b5b0097a8434289fcae00c50af531ed95c4ef0"} Oct 08 00:12:30 crc kubenswrapper[4871]: I1008 00:12:30.606594 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-td4fb" event={"ID":"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793","Type":"ContainerStarted","Data":"44775118be2064bd0ad5d78abd440a9d9f8d0a97a2d3e393a61732904d8bbe15"} Oct 08 00:12:30 crc kubenswrapper[4871]: I1008 00:12:30.608514 4871 generic.go:334] "Generic (PLEG): container finished" podID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerID="dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766" exitCode=0 Oct 08 00:12:30 crc kubenswrapper[4871]: I1008 00:12:30.608543 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mf2vl" event={"ID":"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2","Type":"ContainerDied","Data":"dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766"} Oct 08 00:12:30 crc kubenswrapper[4871]: I1008 00:12:30.633532 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-td4fb" podStartSLOduration=2.621906524 podStartE2EDuration="3.633511251s" podCreationTimestamp="2025-10-08 00:12:27 +0000 UTC" firstStartedPulling="2025-10-08 00:12:28.643887449 +0000 UTC m=+7422.446585522" lastFinishedPulling="2025-10-08 00:12:29.655492176 +0000 UTC m=+7423.458190249" observedRunningTime="2025-10-08 00:12:30.63156878 +0000 UTC m=+7424.434266853" watchObservedRunningTime="2025-10-08 00:12:30.633511251 +0000 UTC m=+7424.436209344" Oct 08 00:12:32 crc kubenswrapper[4871]: I1008 00:12:32.643148 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mf2vl" event={"ID":"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2","Type":"ContainerStarted","Data":"d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e"} Oct 08 00:12:32 crc kubenswrapper[4871]: I1008 00:12:32.673244 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mf2vl" podStartSLOduration=3.704477991 podStartE2EDuration="6.673219838s" podCreationTimestamp="2025-10-08 00:12:26 +0000 UTC" firstStartedPulling="2025-10-08 00:12:28.58865679 +0000 UTC m=+7422.391354863" lastFinishedPulling="2025-10-08 00:12:31.557398597 +0000 UTC m=+7425.360096710" observedRunningTime="2025-10-08 00:12:32.660678469 +0000 UTC m=+7426.463376542" watchObservedRunningTime="2025-10-08 00:12:32.673219838 +0000 UTC m=+7426.475917911" Oct 08 00:12:37 crc kubenswrapper[4871]: I1008 00:12:37.226130 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:37 crc kubenswrapper[4871]: I1008 00:12:37.226764 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:37 crc kubenswrapper[4871]: I1008 00:12:37.283957 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:37 crc kubenswrapper[4871]: I1008 00:12:37.754177 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:37 crc kubenswrapper[4871]: I1008 00:12:37.808221 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mf2vl"] Oct 08 00:12:39 crc kubenswrapper[4871]: I1008 00:12:39.723728 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mf2vl" podUID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerName="registry-server" containerID="cri-o://d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e" gracePeriod=2 Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.241363 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.325572 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-utilities\") pod \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.326941 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-utilities" (OuterVolumeSpecName: "utilities") pod "9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" (UID: "9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.327158 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-catalog-content\") pod \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.331914 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-kube-api-access-pjv5c" (OuterVolumeSpecName: "kube-api-access-pjv5c") pod "9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" (UID: "9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2"). InnerVolumeSpecName "kube-api-access-pjv5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.334185 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjv5c\" (UniqueName: \"kubernetes.io/projected/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-kube-api-access-pjv5c\") pod \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\" (UID: \"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2\") " Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.335363 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjv5c\" (UniqueName: \"kubernetes.io/projected/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-kube-api-access-pjv5c\") on node \"crc\" DevicePath \"\"" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.335391 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.346265 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" (UID: "9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.437545 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.739260 4871 generic.go:334] "Generic (PLEG): container finished" podID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerID="d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e" exitCode=0 Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.739314 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mf2vl" event={"ID":"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2","Type":"ContainerDied","Data":"d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e"} Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.739347 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mf2vl" event={"ID":"9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2","Type":"ContainerDied","Data":"61a9a8e74aa2873a848d271462f30fe487da5d21194e15e9daa9296bb172cdd1"} Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.739373 4871 scope.go:117] "RemoveContainer" containerID="d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.739750 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mf2vl" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.777675 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mf2vl"] Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.778106 4871 scope.go:117] "RemoveContainer" containerID="dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.785270 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mf2vl"] Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.813342 4871 scope.go:117] "RemoveContainer" containerID="d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.853951 4871 scope.go:117] "RemoveContainer" containerID="d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e" Oct 08 00:12:40 crc kubenswrapper[4871]: E1008 00:12:40.854392 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e\": container with ID starting with d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e not found: ID does not exist" containerID="d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.854421 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e"} err="failed to get container status \"d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e\": rpc error: code = NotFound desc = could not find container \"d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e\": container with ID starting with d86363691dbbfd55296773e245d74ddd80ec2efb92af90e5340d6b4a1670282e not found: ID does not exist" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.854441 4871 scope.go:117] "RemoveContainer" containerID="dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766" Oct 08 00:12:40 crc kubenswrapper[4871]: E1008 00:12:40.856163 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766\": container with ID starting with dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766 not found: ID does not exist" containerID="dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.856211 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766"} err="failed to get container status \"dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766\": rpc error: code = NotFound desc = could not find container \"dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766\": container with ID starting with dd6e278232a7a387564df3ee0a6a0044656cffa23253b839bc30f597dc81f766 not found: ID does not exist" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.856252 4871 scope.go:117] "RemoveContainer" containerID="d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad" Oct 08 00:12:40 crc kubenswrapper[4871]: E1008 00:12:40.857253 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad\": container with ID starting with d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad not found: ID does not exist" containerID="d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.857290 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad"} err="failed to get container status \"d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad\": rpc error: code = NotFound desc = could not find container \"d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad\": container with ID starting with d05616b6a68683147e4fa7c3613291c27576f1d10cfe00ff440f88ae6e7405ad not found: ID does not exist" Oct 08 00:12:40 crc kubenswrapper[4871]: I1008 00:12:40.995987 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" path="/var/lib/kubelet/pods/9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2/volumes" Oct 08 00:14:04 crc kubenswrapper[4871]: I1008 00:14:04.740589 4871 generic.go:334] "Generic (PLEG): container finished" podID="0a1b4a8b-4d71-4bd5-b9dc-e784713c2793" containerID="44775118be2064bd0ad5d78abd440a9d9f8d0a97a2d3e393a61732904d8bbe15" exitCode=0 Oct 08 00:14:04 crc kubenswrapper[4871]: I1008 00:14:04.740689 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-td4fb" event={"ID":"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793","Type":"ContainerDied","Data":"44775118be2064bd0ad5d78abd440a9d9f8d0a97a2d3e393a61732904d8bbe15"} Oct 08 00:14:05 crc kubenswrapper[4871]: I1008 00:14:05.512177 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:14:05 crc kubenswrapper[4871]: I1008 00:14:05.512260 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.249929 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.371310 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ceph\") pod \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.371472 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdwqr\" (UniqueName: \"kubernetes.io/projected/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-kube-api-access-vdwqr\") pod \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.371502 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-inventory\") pod \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.372188 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ssh-key\") pod \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\" (UID: \"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793\") " Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.377369 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ceph" (OuterVolumeSpecName: "ceph") pod "0a1b4a8b-4d71-4bd5-b9dc-e784713c2793" (UID: "0a1b4a8b-4d71-4bd5-b9dc-e784713c2793"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.377392 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-kube-api-access-vdwqr" (OuterVolumeSpecName: "kube-api-access-vdwqr") pod "0a1b4a8b-4d71-4bd5-b9dc-e784713c2793" (UID: "0a1b4a8b-4d71-4bd5-b9dc-e784713c2793"). InnerVolumeSpecName "kube-api-access-vdwqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.401568 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-inventory" (OuterVolumeSpecName: "inventory") pod "0a1b4a8b-4d71-4bd5-b9dc-e784713c2793" (UID: "0a1b4a8b-4d71-4bd5-b9dc-e784713c2793"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.402287 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0a1b4a8b-4d71-4bd5-b9dc-e784713c2793" (UID: "0a1b4a8b-4d71-4bd5-b9dc-e784713c2793"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.474698 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.474736 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.474746 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdwqr\" (UniqueName: \"kubernetes.io/projected/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-kube-api-access-vdwqr\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.474760 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a1b4a8b-4d71-4bd5-b9dc-e784713c2793-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.767161 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-td4fb" event={"ID":"0a1b4a8b-4d71-4bd5-b9dc-e784713c2793","Type":"ContainerDied","Data":"325f401a9299f9ef23fe7519b4b5b0097a8434289fcae00c50af531ed95c4ef0"} Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.767204 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="325f401a9299f9ef23fe7519b4b5b0097a8434289fcae00c50af531ed95c4ef0" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.767359 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-td4fb" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.864748 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-4przw"] Oct 08 00:14:06 crc kubenswrapper[4871]: E1008 00:14:06.865471 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerName="extract-utilities" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.865508 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerName="extract-utilities" Oct 08 00:14:06 crc kubenswrapper[4871]: E1008 00:14:06.865528 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a1b4a8b-4d71-4bd5-b9dc-e784713c2793" containerName="download-cache-openstack-openstack-cell1" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.865536 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a1b4a8b-4d71-4bd5-b9dc-e784713c2793" containerName="download-cache-openstack-openstack-cell1" Oct 08 00:14:06 crc kubenswrapper[4871]: E1008 00:14:06.865556 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerName="registry-server" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.865564 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerName="registry-server" Oct 08 00:14:06 crc kubenswrapper[4871]: E1008 00:14:06.865602 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerName="extract-content" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.865611 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerName="extract-content" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.865886 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a1b4a8b-4d71-4bd5-b9dc-e784713c2793" containerName="download-cache-openstack-openstack-cell1" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.865907 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9042340d-1b36-4ce6-8ac3-fb83a8ed1fc2" containerName="registry-server" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.866954 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.869181 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.869943 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.869997 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.870287 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.879024 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-4przw"] Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.996388 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ssh-key\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.996449 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-inventory\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.996588 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d29n4\" (UniqueName: \"kubernetes.io/projected/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-kube-api-access-d29n4\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:06 crc kubenswrapper[4871]: I1008 00:14:06.996634 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ceph\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.098379 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ssh-key\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.098430 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-inventory\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.098563 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d29n4\" (UniqueName: \"kubernetes.io/projected/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-kube-api-access-d29n4\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.098603 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ceph\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.103529 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ceph\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.105117 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ssh-key\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.112251 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-inventory\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.119569 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d29n4\" (UniqueName: \"kubernetes.io/projected/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-kube-api-access-d29n4\") pod \"configure-network-openstack-openstack-cell1-4przw\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.212113 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:14:07 crc kubenswrapper[4871]: I1008 00:14:07.843905 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-4przw"] Oct 08 00:14:08 crc kubenswrapper[4871]: I1008 00:14:08.787914 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-4przw" event={"ID":"f6c907cd-13ef-4d90-be16-aeeb560c2d9e","Type":"ContainerStarted","Data":"59a066ee8b9025a8ec920ff960963688ebca00d265140a04df636e609e4e2e00"} Oct 08 00:14:08 crc kubenswrapper[4871]: I1008 00:14:08.788920 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-4przw" event={"ID":"f6c907cd-13ef-4d90-be16-aeeb560c2d9e","Type":"ContainerStarted","Data":"a0fb40f08f4252deae80a8e1899bfbe6067fe251a89bc922ad2af91702fbf06a"} Oct 08 00:14:08 crc kubenswrapper[4871]: I1008 00:14:08.820153 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-4przw" podStartSLOduration=2.35670911 podStartE2EDuration="2.820126526s" podCreationTimestamp="2025-10-08 00:14:06 +0000 UTC" firstStartedPulling="2025-10-08 00:14:07.851726012 +0000 UTC m=+7521.654424095" lastFinishedPulling="2025-10-08 00:14:08.315143428 +0000 UTC m=+7522.117841511" observedRunningTime="2025-10-08 00:14:08.811242783 +0000 UTC m=+7522.613940886" watchObservedRunningTime="2025-10-08 00:14:08.820126526 +0000 UTC m=+7522.622824629" Oct 08 00:14:35 crc kubenswrapper[4871]: I1008 00:14:35.512157 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:14:35 crc kubenswrapper[4871]: I1008 00:14:35.513007 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.154959 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg"] Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.157587 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.160952 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.161168 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.165005 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg"] Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.166809 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0866e481-6790-4024-b638-0e86221234eb-secret-volume\") pod \"collect-profiles-29331375-5gmrg\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.166945 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0866e481-6790-4024-b638-0e86221234eb-config-volume\") pod \"collect-profiles-29331375-5gmrg\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.167042 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pnk4\" (UniqueName: \"kubernetes.io/projected/0866e481-6790-4024-b638-0e86221234eb-kube-api-access-8pnk4\") pod \"collect-profiles-29331375-5gmrg\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.270069 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pnk4\" (UniqueName: \"kubernetes.io/projected/0866e481-6790-4024-b638-0e86221234eb-kube-api-access-8pnk4\") pod \"collect-profiles-29331375-5gmrg\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.270249 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0866e481-6790-4024-b638-0e86221234eb-secret-volume\") pod \"collect-profiles-29331375-5gmrg\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.270337 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0866e481-6790-4024-b638-0e86221234eb-config-volume\") pod \"collect-profiles-29331375-5gmrg\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.271442 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0866e481-6790-4024-b638-0e86221234eb-config-volume\") pod \"collect-profiles-29331375-5gmrg\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.277753 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0866e481-6790-4024-b638-0e86221234eb-secret-volume\") pod \"collect-profiles-29331375-5gmrg\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.291213 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pnk4\" (UniqueName: \"kubernetes.io/projected/0866e481-6790-4024-b638-0e86221234eb-kube-api-access-8pnk4\") pod \"collect-profiles-29331375-5gmrg\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:00 crc kubenswrapper[4871]: I1008 00:15:00.490647 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:01 crc kubenswrapper[4871]: I1008 00:15:01.002172 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg"] Oct 08 00:15:01 crc kubenswrapper[4871]: I1008 00:15:01.381518 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" event={"ID":"0866e481-6790-4024-b638-0e86221234eb","Type":"ContainerStarted","Data":"9dae4e48919efc401d480c7a543d30e5fcf3a5dbc75164eea3d458358fc6a280"} Oct 08 00:15:01 crc kubenswrapper[4871]: I1008 00:15:01.381564 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" event={"ID":"0866e481-6790-4024-b638-0e86221234eb","Type":"ContainerStarted","Data":"976749555e7ce7f1cf6c7bed29dc5feed77974102887a985a9c7812132c33d3d"} Oct 08 00:15:01 crc kubenswrapper[4871]: I1008 00:15:01.402019 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" podStartSLOduration=1.402000651 podStartE2EDuration="1.402000651s" podCreationTimestamp="2025-10-08 00:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:15:01.394085894 +0000 UTC m=+7575.196783967" watchObservedRunningTime="2025-10-08 00:15:01.402000651 +0000 UTC m=+7575.204698714" Oct 08 00:15:02 crc kubenswrapper[4871]: I1008 00:15:02.391627 4871 generic.go:334] "Generic (PLEG): container finished" podID="0866e481-6790-4024-b638-0e86221234eb" containerID="9dae4e48919efc401d480c7a543d30e5fcf3a5dbc75164eea3d458358fc6a280" exitCode=0 Oct 08 00:15:02 crc kubenswrapper[4871]: I1008 00:15:02.391689 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" event={"ID":"0866e481-6790-4024-b638-0e86221234eb","Type":"ContainerDied","Data":"9dae4e48919efc401d480c7a543d30e5fcf3a5dbc75164eea3d458358fc6a280"} Oct 08 00:15:03 crc kubenswrapper[4871]: I1008 00:15:03.826179 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:03 crc kubenswrapper[4871]: I1008 00:15:03.952242 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0866e481-6790-4024-b638-0e86221234eb-secret-volume\") pod \"0866e481-6790-4024-b638-0e86221234eb\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " Oct 08 00:15:03 crc kubenswrapper[4871]: I1008 00:15:03.952649 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0866e481-6790-4024-b638-0e86221234eb-config-volume\") pod \"0866e481-6790-4024-b638-0e86221234eb\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " Oct 08 00:15:03 crc kubenswrapper[4871]: I1008 00:15:03.952714 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pnk4\" (UniqueName: \"kubernetes.io/projected/0866e481-6790-4024-b638-0e86221234eb-kube-api-access-8pnk4\") pod \"0866e481-6790-4024-b638-0e86221234eb\" (UID: \"0866e481-6790-4024-b638-0e86221234eb\") " Oct 08 00:15:03 crc kubenswrapper[4871]: I1008 00:15:03.953270 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0866e481-6790-4024-b638-0e86221234eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "0866e481-6790-4024-b638-0e86221234eb" (UID: "0866e481-6790-4024-b638-0e86221234eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:15:03 crc kubenswrapper[4871]: I1008 00:15:03.958565 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0866e481-6790-4024-b638-0e86221234eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0866e481-6790-4024-b638-0e86221234eb" (UID: "0866e481-6790-4024-b638-0e86221234eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:15:03 crc kubenswrapper[4871]: I1008 00:15:03.959022 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0866e481-6790-4024-b638-0e86221234eb-kube-api-access-8pnk4" (OuterVolumeSpecName: "kube-api-access-8pnk4") pod "0866e481-6790-4024-b638-0e86221234eb" (UID: "0866e481-6790-4024-b638-0e86221234eb"). InnerVolumeSpecName "kube-api-access-8pnk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:15:04 crc kubenswrapper[4871]: I1008 00:15:04.056411 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0866e481-6790-4024-b638-0e86221234eb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:04 crc kubenswrapper[4871]: I1008 00:15:04.056446 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pnk4\" (UniqueName: \"kubernetes.io/projected/0866e481-6790-4024-b638-0e86221234eb-kube-api-access-8pnk4\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:04 crc kubenswrapper[4871]: I1008 00:15:04.056462 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0866e481-6790-4024-b638-0e86221234eb-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:04 crc kubenswrapper[4871]: I1008 00:15:04.415372 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" event={"ID":"0866e481-6790-4024-b638-0e86221234eb","Type":"ContainerDied","Data":"976749555e7ce7f1cf6c7bed29dc5feed77974102887a985a9c7812132c33d3d"} Oct 08 00:15:04 crc kubenswrapper[4871]: I1008 00:15:04.415429 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="976749555e7ce7f1cf6c7bed29dc5feed77974102887a985a9c7812132c33d3d" Oct 08 00:15:04 crc kubenswrapper[4871]: I1008 00:15:04.415492 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg" Oct 08 00:15:04 crc kubenswrapper[4871]: I1008 00:15:04.496325 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk"] Oct 08 00:15:04 crc kubenswrapper[4871]: I1008 00:15:04.507896 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331330-rmnfk"] Oct 08 00:15:05 crc kubenswrapper[4871]: I1008 00:15:05.002264 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b42486-ae7f-4ef5-b2f9-6adbbce86922" path="/var/lib/kubelet/pods/96b42486-ae7f-4ef5-b2f9-6adbbce86922/volumes" Oct 08 00:15:05 crc kubenswrapper[4871]: I1008 00:15:05.511929 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:15:05 crc kubenswrapper[4871]: I1008 00:15:05.512005 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:15:05 crc kubenswrapper[4871]: I1008 00:15:05.512066 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:15:05 crc kubenswrapper[4871]: I1008 00:15:05.512935 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"032447e2c741f3b9f493330ad578f9e126416d12ba3cc2e1856ed60479a27106"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:15:05 crc kubenswrapper[4871]: I1008 00:15:05.512996 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://032447e2c741f3b9f493330ad578f9e126416d12ba3cc2e1856ed60479a27106" gracePeriod=600 Oct 08 00:15:06 crc kubenswrapper[4871]: I1008 00:15:06.438419 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="032447e2c741f3b9f493330ad578f9e126416d12ba3cc2e1856ed60479a27106" exitCode=0 Oct 08 00:15:06 crc kubenswrapper[4871]: I1008 00:15:06.438492 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"032447e2c741f3b9f493330ad578f9e126416d12ba3cc2e1856ed60479a27106"} Oct 08 00:15:06 crc kubenswrapper[4871]: I1008 00:15:06.439224 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed"} Oct 08 00:15:06 crc kubenswrapper[4871]: I1008 00:15:06.439256 4871 scope.go:117] "RemoveContainer" containerID="02c5aa7eec73d8c958bfe65c93e77234f36b4ac5b26ca22fbd2103b623736f97" Oct 08 00:15:28 crc kubenswrapper[4871]: I1008 00:15:28.680536 4871 generic.go:334] "Generic (PLEG): container finished" podID="f6c907cd-13ef-4d90-be16-aeeb560c2d9e" containerID="59a066ee8b9025a8ec920ff960963688ebca00d265140a04df636e609e4e2e00" exitCode=0 Oct 08 00:15:28 crc kubenswrapper[4871]: I1008 00:15:28.680648 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-4przw" event={"ID":"f6c907cd-13ef-4d90-be16-aeeb560c2d9e","Type":"ContainerDied","Data":"59a066ee8b9025a8ec920ff960963688ebca00d265140a04df636e609e4e2e00"} Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.195708 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.392544 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ssh-key\") pod \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.392688 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ceph\") pod \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.392888 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d29n4\" (UniqueName: \"kubernetes.io/projected/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-kube-api-access-d29n4\") pod \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.392955 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-inventory\") pod \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\" (UID: \"f6c907cd-13ef-4d90-be16-aeeb560c2d9e\") " Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.399653 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ceph" (OuterVolumeSpecName: "ceph") pod "f6c907cd-13ef-4d90-be16-aeeb560c2d9e" (UID: "f6c907cd-13ef-4d90-be16-aeeb560c2d9e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.399874 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-kube-api-access-d29n4" (OuterVolumeSpecName: "kube-api-access-d29n4") pod "f6c907cd-13ef-4d90-be16-aeeb560c2d9e" (UID: "f6c907cd-13ef-4d90-be16-aeeb560c2d9e"). InnerVolumeSpecName "kube-api-access-d29n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.441567 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-inventory" (OuterVolumeSpecName: "inventory") pod "f6c907cd-13ef-4d90-be16-aeeb560c2d9e" (UID: "f6c907cd-13ef-4d90-be16-aeeb560c2d9e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.442935 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f6c907cd-13ef-4d90-be16-aeeb560c2d9e" (UID: "f6c907cd-13ef-4d90-be16-aeeb560c2d9e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.496704 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.496761 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.496783 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d29n4\" (UniqueName: \"kubernetes.io/projected/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-kube-api-access-d29n4\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.496830 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6c907cd-13ef-4d90-be16-aeeb560c2d9e-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.707683 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-4przw" event={"ID":"f6c907cd-13ef-4d90-be16-aeeb560c2d9e","Type":"ContainerDied","Data":"a0fb40f08f4252deae80a8e1899bfbe6067fe251a89bc922ad2af91702fbf06a"} Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.707740 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0fb40f08f4252deae80a8e1899bfbe6067fe251a89bc922ad2af91702fbf06a" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.707761 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-4przw" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.832056 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-96fss"] Oct 08 00:15:30 crc kubenswrapper[4871]: E1008 00:15:30.832735 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0866e481-6790-4024-b638-0e86221234eb" containerName="collect-profiles" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.832764 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="0866e481-6790-4024-b638-0e86221234eb" containerName="collect-profiles" Oct 08 00:15:30 crc kubenswrapper[4871]: E1008 00:15:30.832833 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6c907cd-13ef-4d90-be16-aeeb560c2d9e" containerName="configure-network-openstack-openstack-cell1" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.832852 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6c907cd-13ef-4d90-be16-aeeb560c2d9e" containerName="configure-network-openstack-openstack-cell1" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.833269 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="0866e481-6790-4024-b638-0e86221234eb" containerName="collect-profiles" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.833330 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6c907cd-13ef-4d90-be16-aeeb560c2d9e" containerName="configure-network-openstack-openstack-cell1" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.834740 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.837856 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.838778 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.838820 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.838811 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:15:30 crc kubenswrapper[4871]: I1008 00:15:30.857230 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-96fss"] Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.008225 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ceph\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.008781 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ssh-key\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.008929 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ssbj\" (UniqueName: \"kubernetes.io/projected/404d74ce-e450-47f2-acad-54b000cae498-kube-api-access-2ssbj\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.009129 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-inventory\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.111307 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-inventory\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.111575 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ceph\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.111787 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ssh-key\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.111865 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ssbj\" (UniqueName: \"kubernetes.io/projected/404d74ce-e450-47f2-acad-54b000cae498-kube-api-access-2ssbj\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.118477 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ceph\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.118479 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ssh-key\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.119048 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-inventory\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.137962 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ssbj\" (UniqueName: \"kubernetes.io/projected/404d74ce-e450-47f2-acad-54b000cae498-kube-api-access-2ssbj\") pod \"validate-network-openstack-openstack-cell1-96fss\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.165590 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:31 crc kubenswrapper[4871]: I1008 00:15:31.796131 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-96fss"] Oct 08 00:15:31 crc kubenswrapper[4871]: W1008 00:15:31.802769 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod404d74ce_e450_47f2_acad_54b000cae498.slice/crio-251d9b0d5f113333ccf5f04df348d2cc9d46403518e10ba419a0af55dc734613 WatchSource:0}: Error finding container 251d9b0d5f113333ccf5f04df348d2cc9d46403518e10ba419a0af55dc734613: Status 404 returned error can't find the container with id 251d9b0d5f113333ccf5f04df348d2cc9d46403518e10ba419a0af55dc734613 Oct 08 00:15:32 crc kubenswrapper[4871]: I1008 00:15:32.733040 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-96fss" event={"ID":"404d74ce-e450-47f2-acad-54b000cae498","Type":"ContainerStarted","Data":"49a9be44a5a9e31c972c08a3dbb99413136ebb76536a0d281eca86eb1b402c60"} Oct 08 00:15:32 crc kubenswrapper[4871]: I1008 00:15:32.733645 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-96fss" event={"ID":"404d74ce-e450-47f2-acad-54b000cae498","Type":"ContainerStarted","Data":"251d9b0d5f113333ccf5f04df348d2cc9d46403518e10ba419a0af55dc734613"} Oct 08 00:15:32 crc kubenswrapper[4871]: I1008 00:15:32.765581 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-96fss" podStartSLOduration=2.283725912 podStartE2EDuration="2.765558312s" podCreationTimestamp="2025-10-08 00:15:30 +0000 UTC" firstStartedPulling="2025-10-08 00:15:31.807091039 +0000 UTC m=+7605.609789142" lastFinishedPulling="2025-10-08 00:15:32.288923469 +0000 UTC m=+7606.091621542" observedRunningTime="2025-10-08 00:15:32.755446817 +0000 UTC m=+7606.558144890" watchObservedRunningTime="2025-10-08 00:15:32.765558312 +0000 UTC m=+7606.568256385" Oct 08 00:15:37 crc kubenswrapper[4871]: I1008 00:15:37.790784 4871 generic.go:334] "Generic (PLEG): container finished" podID="404d74ce-e450-47f2-acad-54b000cae498" containerID="49a9be44a5a9e31c972c08a3dbb99413136ebb76536a0d281eca86eb1b402c60" exitCode=0 Oct 08 00:15:37 crc kubenswrapper[4871]: I1008 00:15:37.790836 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-96fss" event={"ID":"404d74ce-e450-47f2-acad-54b000cae498","Type":"ContainerDied","Data":"49a9be44a5a9e31c972c08a3dbb99413136ebb76536a0d281eca86eb1b402c60"} Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.295233 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.326130 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ssh-key\") pod \"404d74ce-e450-47f2-acad-54b000cae498\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.326335 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-inventory\") pod \"404d74ce-e450-47f2-acad-54b000cae498\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.326368 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ceph\") pod \"404d74ce-e450-47f2-acad-54b000cae498\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.326415 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ssbj\" (UniqueName: \"kubernetes.io/projected/404d74ce-e450-47f2-acad-54b000cae498-kube-api-access-2ssbj\") pod \"404d74ce-e450-47f2-acad-54b000cae498\" (UID: \"404d74ce-e450-47f2-acad-54b000cae498\") " Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.332364 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/404d74ce-e450-47f2-acad-54b000cae498-kube-api-access-2ssbj" (OuterVolumeSpecName: "kube-api-access-2ssbj") pod "404d74ce-e450-47f2-acad-54b000cae498" (UID: "404d74ce-e450-47f2-acad-54b000cae498"). InnerVolumeSpecName "kube-api-access-2ssbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.332874 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ceph" (OuterVolumeSpecName: "ceph") pod "404d74ce-e450-47f2-acad-54b000cae498" (UID: "404d74ce-e450-47f2-acad-54b000cae498"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.364051 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-inventory" (OuterVolumeSpecName: "inventory") pod "404d74ce-e450-47f2-acad-54b000cae498" (UID: "404d74ce-e450-47f2-acad-54b000cae498"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.364933 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "404d74ce-e450-47f2-acad-54b000cae498" (UID: "404d74ce-e450-47f2-acad-54b000cae498"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.431344 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.431385 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.431410 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/404d74ce-e450-47f2-acad-54b000cae498-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.431429 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ssbj\" (UniqueName: \"kubernetes.io/projected/404d74ce-e450-47f2-acad-54b000cae498-kube-api-access-2ssbj\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.820462 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-96fss" event={"ID":"404d74ce-e450-47f2-acad-54b000cae498","Type":"ContainerDied","Data":"251d9b0d5f113333ccf5f04df348d2cc9d46403518e10ba419a0af55dc734613"} Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.820524 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="251d9b0d5f113333ccf5f04df348d2cc9d46403518e10ba419a0af55dc734613" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.820601 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-96fss" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.953771 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-pthdq"] Oct 08 00:15:39 crc kubenswrapper[4871]: E1008 00:15:39.954329 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="404d74ce-e450-47f2-acad-54b000cae498" containerName="validate-network-openstack-openstack-cell1" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.954350 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="404d74ce-e450-47f2-acad-54b000cae498" containerName="validate-network-openstack-openstack-cell1" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.954581 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="404d74ce-e450-47f2-acad-54b000cae498" containerName="validate-network-openstack-openstack-cell1" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.955768 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.958135 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.958515 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.958675 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.961282 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:15:39 crc kubenswrapper[4871]: I1008 00:15:39.968214 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-pthdq"] Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.059097 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ceph\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.059185 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ssh-key\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.059461 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-inventory\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.059537 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k22p6\" (UniqueName: \"kubernetes.io/projected/1e32fa4d-959c-4011-8b54-aa7d820ff228-kube-api-access-k22p6\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.161586 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-inventory\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.162060 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k22p6\" (UniqueName: \"kubernetes.io/projected/1e32fa4d-959c-4011-8b54-aa7d820ff228-kube-api-access-k22p6\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.162416 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ceph\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.162451 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ssh-key\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.166944 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-inventory\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.166962 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ceph\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.177733 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ssh-key\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.182807 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k22p6\" (UniqueName: \"kubernetes.io/projected/1e32fa4d-959c-4011-8b54-aa7d820ff228-kube-api-access-k22p6\") pod \"install-os-openstack-openstack-cell1-pthdq\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.279686 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:15:40 crc kubenswrapper[4871]: I1008 00:15:40.881440 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-pthdq"] Oct 08 00:15:41 crc kubenswrapper[4871]: I1008 00:15:41.853285 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-pthdq" event={"ID":"1e32fa4d-959c-4011-8b54-aa7d820ff228","Type":"ContainerStarted","Data":"108a05c6c4c1e38d9ba0dbed859025a28437d8bb173a84266350337c9fa1c710"} Oct 08 00:15:41 crc kubenswrapper[4871]: I1008 00:15:41.853653 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-pthdq" event={"ID":"1e32fa4d-959c-4011-8b54-aa7d820ff228","Type":"ContainerStarted","Data":"011938e6ea2b85bb611faac00c64c85f5ad5dd6ee0a095c94e60c486b7aac6a1"} Oct 08 00:15:41 crc kubenswrapper[4871]: I1008 00:15:41.936001 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-pthdq" podStartSLOduration=2.549982509 podStartE2EDuration="2.935980854s" podCreationTimestamp="2025-10-08 00:15:39 +0000 UTC" firstStartedPulling="2025-10-08 00:15:40.903069258 +0000 UTC m=+7614.705767331" lastFinishedPulling="2025-10-08 00:15:41.289067603 +0000 UTC m=+7615.091765676" observedRunningTime="2025-10-08 00:15:41.902465354 +0000 UTC m=+7615.705163427" watchObservedRunningTime="2025-10-08 00:15:41.935980854 +0000 UTC m=+7615.738678927" Oct 08 00:16:02 crc kubenswrapper[4871]: I1008 00:16:02.787763 4871 scope.go:117] "RemoveContainer" containerID="6119698f5691f10e9cf19cbcb4553b97df2c01a2fed40ea99c4584dc29736092" Oct 08 00:16:27 crc kubenswrapper[4871]: I1008 00:16:27.379040 4871 generic.go:334] "Generic (PLEG): container finished" podID="1e32fa4d-959c-4011-8b54-aa7d820ff228" containerID="108a05c6c4c1e38d9ba0dbed859025a28437d8bb173a84266350337c9fa1c710" exitCode=0 Oct 08 00:16:27 crc kubenswrapper[4871]: I1008 00:16:27.379133 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-pthdq" event={"ID":"1e32fa4d-959c-4011-8b54-aa7d820ff228","Type":"ContainerDied","Data":"108a05c6c4c1e38d9ba0dbed859025a28437d8bb173a84266350337c9fa1c710"} Oct 08 00:16:28 crc kubenswrapper[4871]: I1008 00:16:28.861112 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.029421 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ceph\") pod \"1e32fa4d-959c-4011-8b54-aa7d820ff228\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.029515 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ssh-key\") pod \"1e32fa4d-959c-4011-8b54-aa7d820ff228\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.029718 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k22p6\" (UniqueName: \"kubernetes.io/projected/1e32fa4d-959c-4011-8b54-aa7d820ff228-kube-api-access-k22p6\") pod \"1e32fa4d-959c-4011-8b54-aa7d820ff228\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.029962 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-inventory\") pod \"1e32fa4d-959c-4011-8b54-aa7d820ff228\" (UID: \"1e32fa4d-959c-4011-8b54-aa7d820ff228\") " Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.035303 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e32fa4d-959c-4011-8b54-aa7d820ff228-kube-api-access-k22p6" (OuterVolumeSpecName: "kube-api-access-k22p6") pod "1e32fa4d-959c-4011-8b54-aa7d820ff228" (UID: "1e32fa4d-959c-4011-8b54-aa7d820ff228"). InnerVolumeSpecName "kube-api-access-k22p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.040156 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ceph" (OuterVolumeSpecName: "ceph") pod "1e32fa4d-959c-4011-8b54-aa7d820ff228" (UID: "1e32fa4d-959c-4011-8b54-aa7d820ff228"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.067039 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1e32fa4d-959c-4011-8b54-aa7d820ff228" (UID: "1e32fa4d-959c-4011-8b54-aa7d820ff228"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.082302 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-inventory" (OuterVolumeSpecName: "inventory") pod "1e32fa4d-959c-4011-8b54-aa7d820ff228" (UID: "1e32fa4d-959c-4011-8b54-aa7d820ff228"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.133762 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.134066 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k22p6\" (UniqueName: \"kubernetes.io/projected/1e32fa4d-959c-4011-8b54-aa7d820ff228-kube-api-access-k22p6\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.134091 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.134106 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e32fa4d-959c-4011-8b54-aa7d820ff228-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.409501 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-pthdq" event={"ID":"1e32fa4d-959c-4011-8b54-aa7d820ff228","Type":"ContainerDied","Data":"011938e6ea2b85bb611faac00c64c85f5ad5dd6ee0a095c94e60c486b7aac6a1"} Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.409560 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="011938e6ea2b85bb611faac00c64c85f5ad5dd6ee0a095c94e60c486b7aac6a1" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.409632 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-pthdq" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.494764 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-t9g5h"] Oct 08 00:16:29 crc kubenswrapper[4871]: E1008 00:16:29.495346 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e32fa4d-959c-4011-8b54-aa7d820ff228" containerName="install-os-openstack-openstack-cell1" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.495386 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e32fa4d-959c-4011-8b54-aa7d820ff228" containerName="install-os-openstack-openstack-cell1" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.495672 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e32fa4d-959c-4011-8b54-aa7d820ff228" containerName="install-os-openstack-openstack-cell1" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.496663 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.502487 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.502718 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.504341 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.504607 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.541659 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-t9g5h"] Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.650137 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ceph\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.650236 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qnkm\" (UniqueName: \"kubernetes.io/projected/b587d525-db27-497c-9539-f5492c8e4b33-kube-api-access-8qnkm\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.650303 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ssh-key\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.650454 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-inventory\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.752695 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ceph\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.752869 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qnkm\" (UniqueName: \"kubernetes.io/projected/b587d525-db27-497c-9539-f5492c8e4b33-kube-api-access-8qnkm\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.752956 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ssh-key\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.753004 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-inventory\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.759829 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-inventory\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.761331 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ceph\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.764076 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ssh-key\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.785342 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qnkm\" (UniqueName: \"kubernetes.io/projected/b587d525-db27-497c-9539-f5492c8e4b33-kube-api-access-8qnkm\") pod \"configure-os-openstack-openstack-cell1-t9g5h\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:29 crc kubenswrapper[4871]: I1008 00:16:29.827148 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:16:30 crc kubenswrapper[4871]: W1008 00:16:30.393472 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb587d525_db27_497c_9539_f5492c8e4b33.slice/crio-db38024eac21bf940d0fa39034b7797461c9cd5dc75753bd3df8ba36f71bc8bc WatchSource:0}: Error finding container db38024eac21bf940d0fa39034b7797461c9cd5dc75753bd3df8ba36f71bc8bc: Status 404 returned error can't find the container with id db38024eac21bf940d0fa39034b7797461c9cd5dc75753bd3df8ba36f71bc8bc Oct 08 00:16:30 crc kubenswrapper[4871]: I1008 00:16:30.400506 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-t9g5h"] Oct 08 00:16:30 crc kubenswrapper[4871]: I1008 00:16:30.420980 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" event={"ID":"b587d525-db27-497c-9539-f5492c8e4b33","Type":"ContainerStarted","Data":"db38024eac21bf940d0fa39034b7797461c9cd5dc75753bd3df8ba36f71bc8bc"} Oct 08 00:16:31 crc kubenswrapper[4871]: I1008 00:16:31.435371 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" event={"ID":"b587d525-db27-497c-9539-f5492c8e4b33","Type":"ContainerStarted","Data":"711e2611560af9b5f2fe50f6fd5bb97d22e1baff7de7616502218691aa7627ae"} Oct 08 00:16:31 crc kubenswrapper[4871]: I1008 00:16:31.469200 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" podStartSLOduration=1.959387172 podStartE2EDuration="2.469179225s" podCreationTimestamp="2025-10-08 00:16:29 +0000 UTC" firstStartedPulling="2025-10-08 00:16:30.397471061 +0000 UTC m=+7664.200169134" lastFinishedPulling="2025-10-08 00:16:30.907263104 +0000 UTC m=+7664.709961187" observedRunningTime="2025-10-08 00:16:31.456479662 +0000 UTC m=+7665.259177775" watchObservedRunningTime="2025-10-08 00:16:31.469179225 +0000 UTC m=+7665.271877328" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.237196 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8zcc7"] Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.241942 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.247989 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8zcc7"] Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.330498 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqgsf\" (UniqueName: \"kubernetes.io/projected/5f8d6dfe-9200-4130-ba88-34934aa55ea0-kube-api-access-bqgsf\") pod \"certified-operators-8zcc7\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.330681 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-catalog-content\") pod \"certified-operators-8zcc7\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.330756 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-utilities\") pod \"certified-operators-8zcc7\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.432728 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-catalog-content\") pod \"certified-operators-8zcc7\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.432824 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-utilities\") pod \"certified-operators-8zcc7\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.432903 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqgsf\" (UniqueName: \"kubernetes.io/projected/5f8d6dfe-9200-4130-ba88-34934aa55ea0-kube-api-access-bqgsf\") pod \"certified-operators-8zcc7\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.433577 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-utilities\") pod \"certified-operators-8zcc7\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.433588 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-catalog-content\") pod \"certified-operators-8zcc7\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.456868 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqgsf\" (UniqueName: \"kubernetes.io/projected/5f8d6dfe-9200-4130-ba88-34934aa55ea0-kube-api-access-bqgsf\") pod \"certified-operators-8zcc7\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:02 crc kubenswrapper[4871]: I1008 00:17:02.565566 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:03 crc kubenswrapper[4871]: I1008 00:17:03.129654 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8zcc7"] Oct 08 00:17:03 crc kubenswrapper[4871]: W1008 00:17:03.137481 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f8d6dfe_9200_4130_ba88_34934aa55ea0.slice/crio-d12d78894994ac3f65cc51a562b23aff9ad8b7354ec74367fdaeede577c76dfe WatchSource:0}: Error finding container d12d78894994ac3f65cc51a562b23aff9ad8b7354ec74367fdaeede577c76dfe: Status 404 returned error can't find the container with id d12d78894994ac3f65cc51a562b23aff9ad8b7354ec74367fdaeede577c76dfe Oct 08 00:17:03 crc kubenswrapper[4871]: I1008 00:17:03.778160 4871 generic.go:334] "Generic (PLEG): container finished" podID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerID="9b6989ced397e6a745981a1ccfdedb5d4a15505384edb280e8ff32c7596e551e" exitCode=0 Oct 08 00:17:03 crc kubenswrapper[4871]: I1008 00:17:03.778246 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8zcc7" event={"ID":"5f8d6dfe-9200-4130-ba88-34934aa55ea0","Type":"ContainerDied","Data":"9b6989ced397e6a745981a1ccfdedb5d4a15505384edb280e8ff32c7596e551e"} Oct 08 00:17:03 crc kubenswrapper[4871]: I1008 00:17:03.779632 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8zcc7" event={"ID":"5f8d6dfe-9200-4130-ba88-34934aa55ea0","Type":"ContainerStarted","Data":"d12d78894994ac3f65cc51a562b23aff9ad8b7354ec74367fdaeede577c76dfe"} Oct 08 00:17:05 crc kubenswrapper[4871]: I1008 00:17:05.513848 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:17:05 crc kubenswrapper[4871]: I1008 00:17:05.514144 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:17:08 crc kubenswrapper[4871]: I1008 00:17:08.835652 4871 generic.go:334] "Generic (PLEG): container finished" podID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerID="a492cd44e1f1c4acdcb13e6cb9e34082430ff9e486e71aed68d265c112683042" exitCode=0 Oct 08 00:17:08 crc kubenswrapper[4871]: I1008 00:17:08.835962 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8zcc7" event={"ID":"5f8d6dfe-9200-4130-ba88-34934aa55ea0","Type":"ContainerDied","Data":"a492cd44e1f1c4acdcb13e6cb9e34082430ff9e486e71aed68d265c112683042"} Oct 08 00:17:09 crc kubenswrapper[4871]: I1008 00:17:09.851700 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8zcc7" event={"ID":"5f8d6dfe-9200-4130-ba88-34934aa55ea0","Type":"ContainerStarted","Data":"6808860fccb52642eb89c2255d10654e90788a19449d81fe8a661d7fd1b716b8"} Oct 08 00:17:09 crc kubenswrapper[4871]: I1008 00:17:09.873231 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8zcc7" podStartSLOduration=2.260847956 podStartE2EDuration="7.873210002s" podCreationTimestamp="2025-10-08 00:17:02 +0000 UTC" firstStartedPulling="2025-10-08 00:17:03.78065435 +0000 UTC m=+7697.583352433" lastFinishedPulling="2025-10-08 00:17:09.393016386 +0000 UTC m=+7703.195714479" observedRunningTime="2025-10-08 00:17:09.871301502 +0000 UTC m=+7703.673999595" watchObservedRunningTime="2025-10-08 00:17:09.873210002 +0000 UTC m=+7703.675908075" Oct 08 00:17:12 crc kubenswrapper[4871]: I1008 00:17:12.566593 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:12 crc kubenswrapper[4871]: I1008 00:17:12.567225 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:12 crc kubenswrapper[4871]: I1008 00:17:12.619127 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:17 crc kubenswrapper[4871]: I1008 00:17:17.945384 4871 generic.go:334] "Generic (PLEG): container finished" podID="b587d525-db27-497c-9539-f5492c8e4b33" containerID="711e2611560af9b5f2fe50f6fd5bb97d22e1baff7de7616502218691aa7627ae" exitCode=0 Oct 08 00:17:17 crc kubenswrapper[4871]: I1008 00:17:17.945465 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" event={"ID":"b587d525-db27-497c-9539-f5492c8e4b33","Type":"ContainerDied","Data":"711e2611560af9b5f2fe50f6fd5bb97d22e1baff7de7616502218691aa7627ae"} Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.427596 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.551253 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ceph\") pod \"b587d525-db27-497c-9539-f5492c8e4b33\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.551575 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ssh-key\") pod \"b587d525-db27-497c-9539-f5492c8e4b33\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.551720 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-inventory\") pod \"b587d525-db27-497c-9539-f5492c8e4b33\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.551883 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qnkm\" (UniqueName: \"kubernetes.io/projected/b587d525-db27-497c-9539-f5492c8e4b33-kube-api-access-8qnkm\") pod \"b587d525-db27-497c-9539-f5492c8e4b33\" (UID: \"b587d525-db27-497c-9539-f5492c8e4b33\") " Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.557603 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ceph" (OuterVolumeSpecName: "ceph") pod "b587d525-db27-497c-9539-f5492c8e4b33" (UID: "b587d525-db27-497c-9539-f5492c8e4b33"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.558565 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b587d525-db27-497c-9539-f5492c8e4b33-kube-api-access-8qnkm" (OuterVolumeSpecName: "kube-api-access-8qnkm") pod "b587d525-db27-497c-9539-f5492c8e4b33" (UID: "b587d525-db27-497c-9539-f5492c8e4b33"). InnerVolumeSpecName "kube-api-access-8qnkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.584231 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-inventory" (OuterVolumeSpecName: "inventory") pod "b587d525-db27-497c-9539-f5492c8e4b33" (UID: "b587d525-db27-497c-9539-f5492c8e4b33"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.592446 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b587d525-db27-497c-9539-f5492c8e4b33" (UID: "b587d525-db27-497c-9539-f5492c8e4b33"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.655588 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.655633 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.655650 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qnkm\" (UniqueName: \"kubernetes.io/projected/b587d525-db27-497c-9539-f5492c8e4b33-kube-api-access-8qnkm\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.655665 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b587d525-db27-497c-9539-f5492c8e4b33-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.977902 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" event={"ID":"b587d525-db27-497c-9539-f5492c8e4b33","Type":"ContainerDied","Data":"db38024eac21bf940d0fa39034b7797461c9cd5dc75753bd3df8ba36f71bc8bc"} Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.977960 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db38024eac21bf940d0fa39034b7797461c9cd5dc75753bd3df8ba36f71bc8bc" Oct 08 00:17:19 crc kubenswrapper[4871]: I1008 00:17:19.978047 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-t9g5h" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.084501 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-9w8j5"] Oct 08 00:17:20 crc kubenswrapper[4871]: E1008 00:17:20.085160 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b587d525-db27-497c-9539-f5492c8e4b33" containerName="configure-os-openstack-openstack-cell1" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.085191 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b587d525-db27-497c-9539-f5492c8e4b33" containerName="configure-os-openstack-openstack-cell1" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.085606 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b587d525-db27-497c-9539-f5492c8e4b33" containerName="configure-os-openstack-openstack-cell1" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.086818 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.088665 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.089174 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.089543 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.091148 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.099836 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-9w8j5"] Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.168853 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.169245 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6ll4\" (UniqueName: \"kubernetes.io/projected/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-kube-api-access-s6ll4\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.169424 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ceph\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.169460 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-inventory-0\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.271339 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ceph\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.271382 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-inventory-0\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.271491 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.271530 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6ll4\" (UniqueName: \"kubernetes.io/projected/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-kube-api-access-s6ll4\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.278496 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ceph\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.278992 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.281544 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-inventory-0\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.289266 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6ll4\" (UniqueName: \"kubernetes.io/projected/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-kube-api-access-s6ll4\") pod \"ssh-known-hosts-openstack-9w8j5\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:20 crc kubenswrapper[4871]: I1008 00:17:20.409549 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:21 crc kubenswrapper[4871]: I1008 00:17:21.041981 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-9w8j5"] Oct 08 00:17:22 crc kubenswrapper[4871]: I1008 00:17:22.004022 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-9w8j5" event={"ID":"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0","Type":"ContainerStarted","Data":"ccf2913667d64888d88b08a6292b2d7e189124f22fb0fe5262db387eb329ed56"} Oct 08 00:17:22 crc kubenswrapper[4871]: I1008 00:17:22.004435 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-9w8j5" event={"ID":"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0","Type":"ContainerStarted","Data":"f501d03053cb1c4b058c29cb480be4c5d312208dd295b4f5d204000ef1f4c452"} Oct 08 00:17:22 crc kubenswrapper[4871]: I1008 00:17:22.026110 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-9w8j5" podStartSLOduration=1.393795784 podStartE2EDuration="2.026080751s" podCreationTimestamp="2025-10-08 00:17:20 +0000 UTC" firstStartedPulling="2025-10-08 00:17:21.04749578 +0000 UTC m=+7714.850193853" lastFinishedPulling="2025-10-08 00:17:21.679780747 +0000 UTC m=+7715.482478820" observedRunningTime="2025-10-08 00:17:22.021384678 +0000 UTC m=+7715.824082771" watchObservedRunningTime="2025-10-08 00:17:22.026080751 +0000 UTC m=+7715.828778864" Oct 08 00:17:22 crc kubenswrapper[4871]: I1008 00:17:22.622897 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:17:22 crc kubenswrapper[4871]: I1008 00:17:22.706000 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8zcc7"] Oct 08 00:17:22 crc kubenswrapper[4871]: I1008 00:17:22.772715 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-644z4"] Oct 08 00:17:22 crc kubenswrapper[4871]: I1008 00:17:22.772965 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-644z4" podUID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerName="registry-server" containerID="cri-o://4eaf07e5dfb893f8de9799688857cbf05de76941207b3d75ec95f1c028dd99f4" gracePeriod=2 Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.017066 4871 generic.go:334] "Generic (PLEG): container finished" podID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerID="4eaf07e5dfb893f8de9799688857cbf05de76941207b3d75ec95f1c028dd99f4" exitCode=0 Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.017117 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-644z4" event={"ID":"8a6ef7d8-24dd-4edb-b492-b430adf2b67f","Type":"ContainerDied","Data":"4eaf07e5dfb893f8de9799688857cbf05de76941207b3d75ec95f1c028dd99f4"} Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.353109 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-644z4" Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.459254 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-catalog-content\") pod \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.459349 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-utilities\") pod \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.459454 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgr66\" (UniqueName: \"kubernetes.io/projected/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-kube-api-access-fgr66\") pod \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\" (UID: \"8a6ef7d8-24dd-4edb-b492-b430adf2b67f\") " Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.465450 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-utilities" (OuterVolumeSpecName: "utilities") pod "8a6ef7d8-24dd-4edb-b492-b430adf2b67f" (UID: "8a6ef7d8-24dd-4edb-b492-b430adf2b67f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.469999 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-kube-api-access-fgr66" (OuterVolumeSpecName: "kube-api-access-fgr66") pod "8a6ef7d8-24dd-4edb-b492-b430adf2b67f" (UID: "8a6ef7d8-24dd-4edb-b492-b430adf2b67f"). InnerVolumeSpecName "kube-api-access-fgr66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.538002 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a6ef7d8-24dd-4edb-b492-b430adf2b67f" (UID: "8a6ef7d8-24dd-4edb-b492-b430adf2b67f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.562806 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgr66\" (UniqueName: \"kubernetes.io/projected/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-kube-api-access-fgr66\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.562840 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:23 crc kubenswrapper[4871]: I1008 00:17:23.562866 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ef7d8-24dd-4edb-b492-b430adf2b67f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:24 crc kubenswrapper[4871]: I1008 00:17:24.028238 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-644z4" event={"ID":"8a6ef7d8-24dd-4edb-b492-b430adf2b67f","Type":"ContainerDied","Data":"d4cbef210d33012659a8d50751e2ab4f604426e6888f33eb62959e3c774b45ed"} Oct 08 00:17:24 crc kubenswrapper[4871]: I1008 00:17:24.028302 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-644z4" Oct 08 00:17:24 crc kubenswrapper[4871]: I1008 00:17:24.028561 4871 scope.go:117] "RemoveContainer" containerID="4eaf07e5dfb893f8de9799688857cbf05de76941207b3d75ec95f1c028dd99f4" Oct 08 00:17:24 crc kubenswrapper[4871]: I1008 00:17:24.052952 4871 scope.go:117] "RemoveContainer" containerID="9b45aebedb9fa68e961cc69c307ec232bc2dced405c75f6cda6a0bf7d37e50ea" Oct 08 00:17:24 crc kubenswrapper[4871]: I1008 00:17:24.071658 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-644z4"] Oct 08 00:17:24 crc kubenswrapper[4871]: I1008 00:17:24.079498 4871 scope.go:117] "RemoveContainer" containerID="1df1ebc05b4ef76062421e0f9bd10566b96ca314d967847975c13682218398c0" Oct 08 00:17:24 crc kubenswrapper[4871]: I1008 00:17:24.087617 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-644z4"] Oct 08 00:17:25 crc kubenswrapper[4871]: I1008 00:17:25.031284 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" path="/var/lib/kubelet/pods/8a6ef7d8-24dd-4edb-b492-b430adf2b67f/volumes" Oct 08 00:17:31 crc kubenswrapper[4871]: I1008 00:17:31.125451 4871 generic.go:334] "Generic (PLEG): container finished" podID="d5b4e059-c3a7-41ee-a76f-104e6cfce8a0" containerID="ccf2913667d64888d88b08a6292b2d7e189124f22fb0fe5262db387eb329ed56" exitCode=0 Oct 08 00:17:31 crc kubenswrapper[4871]: I1008 00:17:31.125575 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-9w8j5" event={"ID":"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0","Type":"ContainerDied","Data":"ccf2913667d64888d88b08a6292b2d7e189124f22fb0fe5262db387eb329ed56"} Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.631627 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.674219 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ceph\") pod \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.674427 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-inventory-0\") pod \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.674522 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ssh-key-openstack-cell1\") pod \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.674627 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6ll4\" (UniqueName: \"kubernetes.io/projected/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-kube-api-access-s6ll4\") pod \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\" (UID: \"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0\") " Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.681770 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ceph" (OuterVolumeSpecName: "ceph") pod "d5b4e059-c3a7-41ee-a76f-104e6cfce8a0" (UID: "d5b4e059-c3a7-41ee-a76f-104e6cfce8a0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.685312 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-kube-api-access-s6ll4" (OuterVolumeSpecName: "kube-api-access-s6ll4") pod "d5b4e059-c3a7-41ee-a76f-104e6cfce8a0" (UID: "d5b4e059-c3a7-41ee-a76f-104e6cfce8a0"). InnerVolumeSpecName "kube-api-access-s6ll4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.717285 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d5b4e059-c3a7-41ee-a76f-104e6cfce8a0" (UID: "d5b4e059-c3a7-41ee-a76f-104e6cfce8a0"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.729832 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "d5b4e059-c3a7-41ee-a76f-104e6cfce8a0" (UID: "d5b4e059-c3a7-41ee-a76f-104e6cfce8a0"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.778164 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.778411 4871 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.778483 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:32 crc kubenswrapper[4871]: I1008 00:17:32.778548 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6ll4\" (UniqueName: \"kubernetes.io/projected/d5b4e059-c3a7-41ee-a76f-104e6cfce8a0-kube-api-access-s6ll4\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.149763 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-9w8j5" event={"ID":"d5b4e059-c3a7-41ee-a76f-104e6cfce8a0","Type":"ContainerDied","Data":"f501d03053cb1c4b058c29cb480be4c5d312208dd295b4f5d204000ef1f4c452"} Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.149852 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f501d03053cb1c4b058c29cb480be4c5d312208dd295b4f5d204000ef1f4c452" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.149877 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-9w8j5" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.236073 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-j8zp6"] Oct 08 00:17:33 crc kubenswrapper[4871]: E1008 00:17:33.236737 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerName="extract-content" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.236818 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerName="extract-content" Oct 08 00:17:33 crc kubenswrapper[4871]: E1008 00:17:33.236894 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerName="registry-server" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.236947 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerName="registry-server" Oct 08 00:17:33 crc kubenswrapper[4871]: E1008 00:17:33.237003 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerName="extract-utilities" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.237049 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerName="extract-utilities" Oct 08 00:17:33 crc kubenswrapper[4871]: E1008 00:17:33.237109 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5b4e059-c3a7-41ee-a76f-104e6cfce8a0" containerName="ssh-known-hosts-openstack" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.237153 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5b4e059-c3a7-41ee-a76f-104e6cfce8a0" containerName="ssh-known-hosts-openstack" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.237406 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5b4e059-c3a7-41ee-a76f-104e6cfce8a0" containerName="ssh-known-hosts-openstack" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.237477 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6ef7d8-24dd-4edb-b492-b430adf2b67f" containerName="registry-server" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.238319 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.240446 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.240883 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.241561 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.245217 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.248124 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-j8zp6"] Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.288565 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ssh-key\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.288669 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-inventory\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.288716 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ceph\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.288914 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-488kp\" (UniqueName: \"kubernetes.io/projected/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-kube-api-access-488kp\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.391617 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ssh-key\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.391724 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-inventory\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.391766 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ceph\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.391891 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-488kp\" (UniqueName: \"kubernetes.io/projected/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-kube-api-access-488kp\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.397250 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ssh-key\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.397731 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ceph\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.398315 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-inventory\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.409924 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-488kp\" (UniqueName: \"kubernetes.io/projected/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-kube-api-access-488kp\") pod \"run-os-openstack-openstack-cell1-j8zp6\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:33 crc kubenswrapper[4871]: I1008 00:17:33.565106 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:34 crc kubenswrapper[4871]: I1008 00:17:34.154099 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-j8zp6"] Oct 08 00:17:34 crc kubenswrapper[4871]: I1008 00:17:34.162332 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:17:35 crc kubenswrapper[4871]: I1008 00:17:35.171355 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-j8zp6" event={"ID":"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c","Type":"ContainerStarted","Data":"26f7904fbe68c30f7c761505dca59c3ecdd53baa12bd033bb3beaa6faedf922b"} Oct 08 00:17:35 crc kubenswrapper[4871]: I1008 00:17:35.512117 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:17:35 crc kubenswrapper[4871]: I1008 00:17:35.512607 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:17:36 crc kubenswrapper[4871]: I1008 00:17:36.184996 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-j8zp6" event={"ID":"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c","Type":"ContainerStarted","Data":"3cd69e71f20fd49ddcfebe1f6975c987cb984064089025e5a550212ae6f6d96a"} Oct 08 00:17:36 crc kubenswrapper[4871]: I1008 00:17:36.206186 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-j8zp6" podStartSLOduration=2.495946628 podStartE2EDuration="3.206168338s" podCreationTimestamp="2025-10-08 00:17:33 +0000 UTC" firstStartedPulling="2025-10-08 00:17:34.161761299 +0000 UTC m=+7727.964459422" lastFinishedPulling="2025-10-08 00:17:34.871983059 +0000 UTC m=+7728.674681132" observedRunningTime="2025-10-08 00:17:36.204112434 +0000 UTC m=+7730.006810517" watchObservedRunningTime="2025-10-08 00:17:36.206168338 +0000 UTC m=+7730.008866411" Oct 08 00:17:44 crc kubenswrapper[4871]: I1008 00:17:44.269318 4871 generic.go:334] "Generic (PLEG): container finished" podID="f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c" containerID="3cd69e71f20fd49ddcfebe1f6975c987cb984064089025e5a550212ae6f6d96a" exitCode=0 Oct 08 00:17:44 crc kubenswrapper[4871]: I1008 00:17:44.269427 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-j8zp6" event={"ID":"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c","Type":"ContainerDied","Data":"3cd69e71f20fd49ddcfebe1f6975c987cb984064089025e5a550212ae6f6d96a"} Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.758752 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.790715 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ssh-key\") pod \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.790896 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-488kp\" (UniqueName: \"kubernetes.io/projected/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-kube-api-access-488kp\") pod \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.790922 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-inventory\") pod \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.791032 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ceph\") pod \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\" (UID: \"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c\") " Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.798007 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-kube-api-access-488kp" (OuterVolumeSpecName: "kube-api-access-488kp") pod "f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c" (UID: "f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c"). InnerVolumeSpecName "kube-api-access-488kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.800913 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ceph" (OuterVolumeSpecName: "ceph") pod "f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c" (UID: "f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.821730 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c" (UID: "f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.834738 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-inventory" (OuterVolumeSpecName: "inventory") pod "f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c" (UID: "f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.893944 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.893990 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-488kp\" (UniqueName: \"kubernetes.io/projected/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-kube-api-access-488kp\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.894006 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:45 crc kubenswrapper[4871]: I1008 00:17:45.894018 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.296928 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-j8zp6" event={"ID":"f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c","Type":"ContainerDied","Data":"26f7904fbe68c30f7c761505dca59c3ecdd53baa12bd033bb3beaa6faedf922b"} Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.296980 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26f7904fbe68c30f7c761505dca59c3ecdd53baa12bd033bb3beaa6faedf922b" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.297062 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-j8zp6" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.366718 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mj2bz"] Oct 08 00:17:46 crc kubenswrapper[4871]: E1008 00:17:46.367461 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c" containerName="run-os-openstack-openstack-cell1" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.367485 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c" containerName="run-os-openstack-openstack-cell1" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.367720 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c" containerName="run-os-openstack-openstack-cell1" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.369597 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.371706 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.371908 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.372044 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.372124 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.377523 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mj2bz"] Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.405292 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ceph\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.405388 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p56xm\" (UniqueName: \"kubernetes.io/projected/88906f81-9b44-4372-9962-6291c10fdb57-kube-api-access-p56xm\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.405978 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.406039 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-inventory\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.507483 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.507542 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-inventory\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.507657 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ceph\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.507680 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p56xm\" (UniqueName: \"kubernetes.io/projected/88906f81-9b44-4372-9962-6291c10fdb57-kube-api-access-p56xm\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.512282 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.512317 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-inventory\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.512738 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ceph\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.525616 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p56xm\" (UniqueName: \"kubernetes.io/projected/88906f81-9b44-4372-9962-6291c10fdb57-kube-api-access-p56xm\") pod \"reboot-os-openstack-openstack-cell1-mj2bz\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:46 crc kubenswrapper[4871]: I1008 00:17:46.706361 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:17:47 crc kubenswrapper[4871]: W1008 00:17:47.283773 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88906f81_9b44_4372_9962_6291c10fdb57.slice/crio-d2a1661f307b7d30f1a0092555fffc07bcd167043758d4c89720341d522e00a9 WatchSource:0}: Error finding container d2a1661f307b7d30f1a0092555fffc07bcd167043758d4c89720341d522e00a9: Status 404 returned error can't find the container with id d2a1661f307b7d30f1a0092555fffc07bcd167043758d4c89720341d522e00a9 Oct 08 00:17:47 crc kubenswrapper[4871]: I1008 00:17:47.284427 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mj2bz"] Oct 08 00:17:47 crc kubenswrapper[4871]: I1008 00:17:47.313138 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" event={"ID":"88906f81-9b44-4372-9962-6291c10fdb57","Type":"ContainerStarted","Data":"d2a1661f307b7d30f1a0092555fffc07bcd167043758d4c89720341d522e00a9"} Oct 08 00:17:47 crc kubenswrapper[4871]: I1008 00:17:47.848501 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:17:48 crc kubenswrapper[4871]: I1008 00:17:48.331759 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" event={"ID":"88906f81-9b44-4372-9962-6291c10fdb57","Type":"ContainerStarted","Data":"eacfc661661bef4a46e03e8dfb54b57753ef5128e92e1a1bfea90eb7a7995a69"} Oct 08 00:17:48 crc kubenswrapper[4871]: I1008 00:17:48.356063 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" podStartSLOduration=1.798759488 podStartE2EDuration="2.356042747s" podCreationTimestamp="2025-10-08 00:17:46 +0000 UTC" firstStartedPulling="2025-10-08 00:17:47.287051135 +0000 UTC m=+7741.089749228" lastFinishedPulling="2025-10-08 00:17:47.844334394 +0000 UTC m=+7741.647032487" observedRunningTime="2025-10-08 00:17:48.350649136 +0000 UTC m=+7742.153347229" watchObservedRunningTime="2025-10-08 00:17:48.356042747 +0000 UTC m=+7742.158740830" Oct 08 00:18:04 crc kubenswrapper[4871]: I1008 00:18:04.513767 4871 generic.go:334] "Generic (PLEG): container finished" podID="88906f81-9b44-4372-9962-6291c10fdb57" containerID="eacfc661661bef4a46e03e8dfb54b57753ef5128e92e1a1bfea90eb7a7995a69" exitCode=0 Oct 08 00:18:04 crc kubenswrapper[4871]: I1008 00:18:04.513955 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" event={"ID":"88906f81-9b44-4372-9962-6291c10fdb57","Type":"ContainerDied","Data":"eacfc661661bef4a46e03e8dfb54b57753ef5128e92e1a1bfea90eb7a7995a69"} Oct 08 00:18:05 crc kubenswrapper[4871]: I1008 00:18:05.512732 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:18:05 crc kubenswrapper[4871]: I1008 00:18:05.512825 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:18:05 crc kubenswrapper[4871]: I1008 00:18:05.512884 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:18:05 crc kubenswrapper[4871]: I1008 00:18:05.513881 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:18:05 crc kubenswrapper[4871]: I1008 00:18:05.513956 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" gracePeriod=600 Oct 08 00:18:05 crc kubenswrapper[4871]: E1008 00:18:05.655228 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.035591 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.115159 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p56xm\" (UniqueName: \"kubernetes.io/projected/88906f81-9b44-4372-9962-6291c10fdb57-kube-api-access-p56xm\") pod \"88906f81-9b44-4372-9962-6291c10fdb57\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.115224 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ssh-key\") pod \"88906f81-9b44-4372-9962-6291c10fdb57\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.115349 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-inventory\") pod \"88906f81-9b44-4372-9962-6291c10fdb57\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.115387 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ceph\") pod \"88906f81-9b44-4372-9962-6291c10fdb57\" (UID: \"88906f81-9b44-4372-9962-6291c10fdb57\") " Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.121945 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ceph" (OuterVolumeSpecName: "ceph") pod "88906f81-9b44-4372-9962-6291c10fdb57" (UID: "88906f81-9b44-4372-9962-6291c10fdb57"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.122048 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88906f81-9b44-4372-9962-6291c10fdb57-kube-api-access-p56xm" (OuterVolumeSpecName: "kube-api-access-p56xm") pod "88906f81-9b44-4372-9962-6291c10fdb57" (UID: "88906f81-9b44-4372-9962-6291c10fdb57"). InnerVolumeSpecName "kube-api-access-p56xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.156640 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-inventory" (OuterVolumeSpecName: "inventory") pod "88906f81-9b44-4372-9962-6291c10fdb57" (UID: "88906f81-9b44-4372-9962-6291c10fdb57"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.156938 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "88906f81-9b44-4372-9962-6291c10fdb57" (UID: "88906f81-9b44-4372-9962-6291c10fdb57"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.219784 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.219888 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.219905 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p56xm\" (UniqueName: \"kubernetes.io/projected/88906f81-9b44-4372-9962-6291c10fdb57-kube-api-access-p56xm\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.219924 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88906f81-9b44-4372-9962-6291c10fdb57-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.535185 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" event={"ID":"88906f81-9b44-4372-9962-6291c10fdb57","Type":"ContainerDied","Data":"d2a1661f307b7d30f1a0092555fffc07bcd167043758d4c89720341d522e00a9"} Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.535499 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2a1661f307b7d30f1a0092555fffc07bcd167043758d4c89720341d522e00a9" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.535252 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mj2bz" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.539173 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" exitCode=0 Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.539226 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed"} Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.539267 4871 scope.go:117] "RemoveContainer" containerID="032447e2c741f3b9f493330ad578f9e126416d12ba3cc2e1856ed60479a27106" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.540338 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:18:06 crc kubenswrapper[4871]: E1008 00:18:06.540934 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.643512 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-tsvxx"] Oct 08 00:18:06 crc kubenswrapper[4871]: E1008 00:18:06.645452 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88906f81-9b44-4372-9962-6291c10fdb57" containerName="reboot-os-openstack-openstack-cell1" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.645604 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="88906f81-9b44-4372-9962-6291c10fdb57" containerName="reboot-os-openstack-openstack-cell1" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.646182 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="88906f81-9b44-4372-9962-6291c10fdb57" containerName="reboot-os-openstack-openstack-cell1" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.647449 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.655896 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.656035 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.656038 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.656725 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.663843 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-tsvxx"] Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.728747 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ssh-key\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.728918 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.728967 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.729265 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.729547 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ceph\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.729723 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.729934 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.730051 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.730084 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.730239 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvtdb\" (UniqueName: \"kubernetes.io/projected/24161d8f-3e5c-4225-aecf-89a7d3d825cc-kube-api-access-vvtdb\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.730321 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.730361 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-inventory\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.832880 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.833186 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ceph\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.833276 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.833367 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.833497 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.833609 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.833768 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvtdb\" (UniqueName: \"kubernetes.io/projected/24161d8f-3e5c-4225-aecf-89a7d3d825cc-kube-api-access-vvtdb\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.834217 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.834691 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-inventory\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.834971 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ssh-key\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.835113 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.835268 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.838761 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.839359 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-inventory\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.839986 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ssh-key\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.840062 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.840966 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.840986 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ceph\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.841044 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.841178 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.841735 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.841915 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.843689 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.861267 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvtdb\" (UniqueName: \"kubernetes.io/projected/24161d8f-3e5c-4225-aecf-89a7d3d825cc-kube-api-access-vvtdb\") pod \"install-certs-openstack-openstack-cell1-tsvxx\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:06 crc kubenswrapper[4871]: I1008 00:18:06.975034 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:07 crc kubenswrapper[4871]: I1008 00:18:07.544173 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-tsvxx"] Oct 08 00:18:08 crc kubenswrapper[4871]: I1008 00:18:08.567138 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" event={"ID":"24161d8f-3e5c-4225-aecf-89a7d3d825cc","Type":"ContainerStarted","Data":"e7a91c15bc07a3a89d362508e5ca8f015bdb61f831c44c5e651a7b2dd941938c"} Oct 08 00:18:08 crc kubenswrapper[4871]: I1008 00:18:08.568498 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" event={"ID":"24161d8f-3e5c-4225-aecf-89a7d3d825cc","Type":"ContainerStarted","Data":"8db62c77f1ae332fbfa552e3fe974fcefbc17119154405fb9173931d993cad2b"} Oct 08 00:18:08 crc kubenswrapper[4871]: I1008 00:18:08.596657 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" podStartSLOduration=1.9908137030000002 podStartE2EDuration="2.596639245s" podCreationTimestamp="2025-10-08 00:18:06 +0000 UTC" firstStartedPulling="2025-10-08 00:18:07.545181653 +0000 UTC m=+7761.347879736" lastFinishedPulling="2025-10-08 00:18:08.151007205 +0000 UTC m=+7761.953705278" observedRunningTime="2025-10-08 00:18:08.587753812 +0000 UTC m=+7762.390451875" watchObservedRunningTime="2025-10-08 00:18:08.596639245 +0000 UTC m=+7762.399337318" Oct 08 00:18:18 crc kubenswrapper[4871]: I1008 00:18:18.982220 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:18:18 crc kubenswrapper[4871]: E1008 00:18:18.983092 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:18:27 crc kubenswrapper[4871]: I1008 00:18:27.794484 4871 generic.go:334] "Generic (PLEG): container finished" podID="24161d8f-3e5c-4225-aecf-89a7d3d825cc" containerID="e7a91c15bc07a3a89d362508e5ca8f015bdb61f831c44c5e651a7b2dd941938c" exitCode=0 Oct 08 00:18:27 crc kubenswrapper[4871]: I1008 00:18:27.794579 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" event={"ID":"24161d8f-3e5c-4225-aecf-89a7d3d825cc","Type":"ContainerDied","Data":"e7a91c15bc07a3a89d362508e5ca8f015bdb61f831c44c5e651a7b2dd941938c"} Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.302738 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472416 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ssh-key\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472515 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ceph\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472570 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-dhcp-combined-ca-bundle\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472612 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-metadata-combined-ca-bundle\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472678 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-nova-combined-ca-bundle\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472702 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ovn-combined-ca-bundle\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472726 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvtdb\" (UniqueName: \"kubernetes.io/projected/24161d8f-3e5c-4225-aecf-89a7d3d825cc-kube-api-access-vvtdb\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472863 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-bootstrap-combined-ca-bundle\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472911 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-libvirt-combined-ca-bundle\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472944 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-sriov-combined-ca-bundle\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.472972 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-telemetry-combined-ca-bundle\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.473024 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-inventory\") pod \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\" (UID: \"24161d8f-3e5c-4225-aecf-89a7d3d825cc\") " Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.479139 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.479183 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.479287 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.480668 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.480770 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.481422 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.482841 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24161d8f-3e5c-4225-aecf-89a7d3d825cc-kube-api-access-vvtdb" (OuterVolumeSpecName: "kube-api-access-vvtdb") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "kube-api-access-vvtdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.482841 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.482924 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ceph" (OuterVolumeSpecName: "ceph") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.485686 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.513965 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-inventory" (OuterVolumeSpecName: "inventory") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.516560 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "24161d8f-3e5c-4225-aecf-89a7d3d825cc" (UID: "24161d8f-3e5c-4225-aecf-89a7d3d825cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576484 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576562 4871 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576588 4871 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576609 4871 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576630 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576648 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvtdb\" (UniqueName: \"kubernetes.io/projected/24161d8f-3e5c-4225-aecf-89a7d3d825cc-kube-api-access-vvtdb\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576665 4871 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576683 4871 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576702 4871 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576719 4871 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576737 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.576753 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24161d8f-3e5c-4225-aecf-89a7d3d825cc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.818764 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" event={"ID":"24161d8f-3e5c-4225-aecf-89a7d3d825cc","Type":"ContainerDied","Data":"8db62c77f1ae332fbfa552e3fe974fcefbc17119154405fb9173931d993cad2b"} Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.818898 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8db62c77f1ae332fbfa552e3fe974fcefbc17119154405fb9173931d993cad2b" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.819001 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-tsvxx" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.920902 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-j98xd"] Oct 08 00:18:29 crc kubenswrapper[4871]: E1008 00:18:29.921930 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24161d8f-3e5c-4225-aecf-89a7d3d825cc" containerName="install-certs-openstack-openstack-cell1" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.921954 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="24161d8f-3e5c-4225-aecf-89a7d3d825cc" containerName="install-certs-openstack-openstack-cell1" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.922311 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="24161d8f-3e5c-4225-aecf-89a7d3d825cc" containerName="install-certs-openstack-openstack-cell1" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.926778 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.929333 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.931098 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.931487 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.932123 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:18:29 crc kubenswrapper[4871]: I1008 00:18:29.940390 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-j98xd"] Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.086979 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-inventory\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.087024 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.087135 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsxhx\" (UniqueName: \"kubernetes.io/projected/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-kube-api-access-qsxhx\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.087197 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ceph\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.189648 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ceph\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.190000 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-inventory\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.190052 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.191127 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsxhx\" (UniqueName: \"kubernetes.io/projected/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-kube-api-access-qsxhx\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.195243 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.195271 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-inventory\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.199916 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ceph\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.212305 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsxhx\" (UniqueName: \"kubernetes.io/projected/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-kube-api-access-qsxhx\") pod \"ceph-client-openstack-openstack-cell1-j98xd\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.258486 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.816127 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-j98xd"] Oct 08 00:18:30 crc kubenswrapper[4871]: W1008 00:18:30.818581 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04f0cc94_5c7f_4cfd_94ea_f0e9556f4cc6.slice/crio-1eb62f51642cff40c9cb8b3c99cbd131e211d36b934ba895f9d9b397a6ddfdab WatchSource:0}: Error finding container 1eb62f51642cff40c9cb8b3c99cbd131e211d36b934ba895f9d9b397a6ddfdab: Status 404 returned error can't find the container with id 1eb62f51642cff40c9cb8b3c99cbd131e211d36b934ba895f9d9b397a6ddfdab Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.829974 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" event={"ID":"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6","Type":"ContainerStarted","Data":"1eb62f51642cff40c9cb8b3c99cbd131e211d36b934ba895f9d9b397a6ddfdab"} Oct 08 00:18:30 crc kubenswrapper[4871]: I1008 00:18:30.984732 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:18:30 crc kubenswrapper[4871]: E1008 00:18:30.985149 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:18:31 crc kubenswrapper[4871]: I1008 00:18:31.845189 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" event={"ID":"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6","Type":"ContainerStarted","Data":"487cf29677dfeb70d108477292771ba27fabd48a57fba100deb82d780c819cd0"} Oct 08 00:18:31 crc kubenswrapper[4871]: I1008 00:18:31.875288 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" podStartSLOduration=2.108037351 podStartE2EDuration="2.875267427s" podCreationTimestamp="2025-10-08 00:18:29 +0000 UTC" firstStartedPulling="2025-10-08 00:18:30.822484751 +0000 UTC m=+7784.625182824" lastFinishedPulling="2025-10-08 00:18:31.589714807 +0000 UTC m=+7785.392412900" observedRunningTime="2025-10-08 00:18:31.861829185 +0000 UTC m=+7785.664527268" watchObservedRunningTime="2025-10-08 00:18:31.875267427 +0000 UTC m=+7785.677965500" Oct 08 00:18:36 crc kubenswrapper[4871]: I1008 00:18:36.916142 4871 generic.go:334] "Generic (PLEG): container finished" podID="04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6" containerID="487cf29677dfeb70d108477292771ba27fabd48a57fba100deb82d780c819cd0" exitCode=0 Oct 08 00:18:36 crc kubenswrapper[4871]: I1008 00:18:36.916232 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" event={"ID":"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6","Type":"ContainerDied","Data":"487cf29677dfeb70d108477292771ba27fabd48a57fba100deb82d780c819cd0"} Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.458258 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.608273 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsxhx\" (UniqueName: \"kubernetes.io/projected/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-kube-api-access-qsxhx\") pod \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.608348 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ssh-key\") pod \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.608396 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ceph\") pod \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.608557 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-inventory\") pod \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\" (UID: \"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6\") " Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.615009 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-kube-api-access-qsxhx" (OuterVolumeSpecName: "kube-api-access-qsxhx") pod "04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6" (UID: "04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6"). InnerVolumeSpecName "kube-api-access-qsxhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.618079 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ceph" (OuterVolumeSpecName: "ceph") pod "04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6" (UID: "04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.639923 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6" (UID: "04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.658603 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-inventory" (OuterVolumeSpecName: "inventory") pod "04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6" (UID: "04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.713179 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsxhx\" (UniqueName: \"kubernetes.io/projected/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-kube-api-access-qsxhx\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.713232 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.713251 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.713269 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.959620 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" event={"ID":"04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6","Type":"ContainerDied","Data":"1eb62f51642cff40c9cb8b3c99cbd131e211d36b934ba895f9d9b397a6ddfdab"} Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.959688 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1eb62f51642cff40c9cb8b3c99cbd131e211d36b934ba895f9d9b397a6ddfdab" Oct 08 00:18:38 crc kubenswrapper[4871]: I1008 00:18:38.959828 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-j98xd" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.038752 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-26h6r"] Oct 08 00:18:39 crc kubenswrapper[4871]: E1008 00:18:39.039421 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6" containerName="ceph-client-openstack-openstack-cell1" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.039442 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6" containerName="ceph-client-openstack-openstack-cell1" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.039856 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6" containerName="ceph-client-openstack-openstack-cell1" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.040895 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.046157 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.049737 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-26h6r"] Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.055647 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.056010 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.056379 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.057651 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.122967 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.123022 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-inventory\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.123069 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.123132 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ceph\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.123244 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ssh-key\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.123285 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9w2t\" (UniqueName: \"kubernetes.io/projected/9efec3ad-9ec3-4008-88b7-fb1c505d459b-kube-api-access-z9w2t\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.225390 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.225685 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-inventory\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.225836 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.226638 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.227158 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ceph\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.227425 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ssh-key\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.227512 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9w2t\" (UniqueName: \"kubernetes.io/projected/9efec3ad-9ec3-4008-88b7-fb1c505d459b-kube-api-access-z9w2t\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.232543 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-inventory\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.232608 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ssh-key\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.232726 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.232770 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ceph\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.244321 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9w2t\" (UniqueName: \"kubernetes.io/projected/9efec3ad-9ec3-4008-88b7-fb1c505d459b-kube-api-access-z9w2t\") pod \"ovn-openstack-openstack-cell1-26h6r\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.367004 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.962459 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-26h6r"] Oct 08 00:18:39 crc kubenswrapper[4871]: I1008 00:18:39.986988 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-26h6r" event={"ID":"9efec3ad-9ec3-4008-88b7-fb1c505d459b","Type":"ContainerStarted","Data":"14f1f6e441bd716107482a3be3d1fa0905b9e971d4a7651dfb61732de1add28c"} Oct 08 00:18:42 crc kubenswrapper[4871]: I1008 00:18:42.007850 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-26h6r" event={"ID":"9efec3ad-9ec3-4008-88b7-fb1c505d459b","Type":"ContainerStarted","Data":"7fc2af2248476bbdf4e71ba2b5cb34786831c213640c5dcfa00e03d5e703b1bb"} Oct 08 00:18:42 crc kubenswrapper[4871]: I1008 00:18:42.038547 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-26h6r" podStartSLOduration=2.243086538 podStartE2EDuration="3.038515664s" podCreationTimestamp="2025-10-08 00:18:39 +0000 UTC" firstStartedPulling="2025-10-08 00:18:39.952734349 +0000 UTC m=+7793.755432432" lastFinishedPulling="2025-10-08 00:18:40.748163445 +0000 UTC m=+7794.550861558" observedRunningTime="2025-10-08 00:18:42.028077681 +0000 UTC m=+7795.830775774" watchObservedRunningTime="2025-10-08 00:18:42.038515664 +0000 UTC m=+7795.841213777" Oct 08 00:18:42 crc kubenswrapper[4871]: I1008 00:18:42.984029 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:18:42 crc kubenswrapper[4871]: E1008 00:18:42.985286 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:18:53 crc kubenswrapper[4871]: I1008 00:18:53.983201 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:18:53 crc kubenswrapper[4871]: E1008 00:18:53.983967 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:19:07 crc kubenswrapper[4871]: I1008 00:19:07.983281 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:19:07 crc kubenswrapper[4871]: E1008 00:19:07.984653 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:19:21 crc kubenswrapper[4871]: I1008 00:19:21.983457 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:19:21 crc kubenswrapper[4871]: E1008 00:19:21.984215 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:19:35 crc kubenswrapper[4871]: I1008 00:19:35.982123 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:19:35 crc kubenswrapper[4871]: E1008 00:19:35.982940 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:19:48 crc kubenswrapper[4871]: I1008 00:19:48.982810 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:19:48 crc kubenswrapper[4871]: E1008 00:19:48.983756 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:19:49 crc kubenswrapper[4871]: I1008 00:19:49.746165 4871 generic.go:334] "Generic (PLEG): container finished" podID="9efec3ad-9ec3-4008-88b7-fb1c505d459b" containerID="7fc2af2248476bbdf4e71ba2b5cb34786831c213640c5dcfa00e03d5e703b1bb" exitCode=0 Oct 08 00:19:49 crc kubenswrapper[4871]: I1008 00:19:49.746295 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-26h6r" event={"ID":"9efec3ad-9ec3-4008-88b7-fb1c505d459b","Type":"ContainerDied","Data":"7fc2af2248476bbdf4e71ba2b5cb34786831c213640c5dcfa00e03d5e703b1bb"} Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.331384 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.467732 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-inventory\") pod \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.467820 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ssh-key\") pod \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.467842 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovncontroller-config-0\") pod \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.468053 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9w2t\" (UniqueName: \"kubernetes.io/projected/9efec3ad-9ec3-4008-88b7-fb1c505d459b-kube-api-access-z9w2t\") pod \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.468114 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ceph\") pod \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.468154 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovn-combined-ca-bundle\") pod \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\" (UID: \"9efec3ad-9ec3-4008-88b7-fb1c505d459b\") " Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.478329 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9efec3ad-9ec3-4008-88b7-fb1c505d459b-kube-api-access-z9w2t" (OuterVolumeSpecName: "kube-api-access-z9w2t") pod "9efec3ad-9ec3-4008-88b7-fb1c505d459b" (UID: "9efec3ad-9ec3-4008-88b7-fb1c505d459b"). InnerVolumeSpecName "kube-api-access-z9w2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.482901 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "9efec3ad-9ec3-4008-88b7-fb1c505d459b" (UID: "9efec3ad-9ec3-4008-88b7-fb1c505d459b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.483003 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ceph" (OuterVolumeSpecName: "ceph") pod "9efec3ad-9ec3-4008-88b7-fb1c505d459b" (UID: "9efec3ad-9ec3-4008-88b7-fb1c505d459b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.495000 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "9efec3ad-9ec3-4008-88b7-fb1c505d459b" (UID: "9efec3ad-9ec3-4008-88b7-fb1c505d459b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.497974 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-inventory" (OuterVolumeSpecName: "inventory") pod "9efec3ad-9ec3-4008-88b7-fb1c505d459b" (UID: "9efec3ad-9ec3-4008-88b7-fb1c505d459b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.501930 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9efec3ad-9ec3-4008-88b7-fb1c505d459b" (UID: "9efec3ad-9ec3-4008-88b7-fb1c505d459b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.571231 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9w2t\" (UniqueName: \"kubernetes.io/projected/9efec3ad-9ec3-4008-88b7-fb1c505d459b-kube-api-access-z9w2t\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.571400 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.571455 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.571505 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.571574 4871 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.571624 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9efec3ad-9ec3-4008-88b7-fb1c505d459b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.776708 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-26h6r" event={"ID":"9efec3ad-9ec3-4008-88b7-fb1c505d459b","Type":"ContainerDied","Data":"14f1f6e441bd716107482a3be3d1fa0905b9e971d4a7651dfb61732de1add28c"} Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.776765 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14f1f6e441bd716107482a3be3d1fa0905b9e971d4a7651dfb61732de1add28c" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.776866 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-26h6r" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.950039 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-qnn9h"] Oct 08 00:19:51 crc kubenswrapper[4871]: E1008 00:19:51.950712 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9efec3ad-9ec3-4008-88b7-fb1c505d459b" containerName="ovn-openstack-openstack-cell1" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.950741 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="9efec3ad-9ec3-4008-88b7-fb1c505d459b" containerName="ovn-openstack-openstack-cell1" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.951085 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="9efec3ad-9ec3-4008-88b7-fb1c505d459b" containerName="ovn-openstack-openstack-cell1" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.952464 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.955302 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.956593 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.956882 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.956892 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.957074 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.957277 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:19:51 crc kubenswrapper[4871]: I1008 00:19:51.959476 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-qnn9h"] Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.086703 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.087063 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.087203 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk6l6\" (UniqueName: \"kubernetes.io/projected/e2f9a2a7-995e-46a7-9272-aa54c93afa60-kube-api-access-lk6l6\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.087341 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.087518 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.087675 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.087821 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.189996 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.190494 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.190645 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk6l6\" (UniqueName: \"kubernetes.io/projected/e2f9a2a7-995e-46a7-9272-aa54c93afa60-kube-api-access-lk6l6\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.190727 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.190880 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.190965 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.191022 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.194542 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.194745 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.196931 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.196935 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.197300 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.207485 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.208816 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk6l6\" (UniqueName: \"kubernetes.io/projected/e2f9a2a7-995e-46a7-9272-aa54c93afa60-kube-api-access-lk6l6\") pod \"neutron-metadata-openstack-openstack-cell1-qnn9h\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.290216 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:19:52 crc kubenswrapper[4871]: I1008 00:19:52.893867 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-qnn9h"] Oct 08 00:19:52 crc kubenswrapper[4871]: W1008 00:19:52.905927 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2f9a2a7_995e_46a7_9272_aa54c93afa60.slice/crio-d127f4f019be18969542bb4f9b890da62ee47ac1691f3ac906d4837206377765 WatchSource:0}: Error finding container d127f4f019be18969542bb4f9b890da62ee47ac1691f3ac906d4837206377765: Status 404 returned error can't find the container with id d127f4f019be18969542bb4f9b890da62ee47ac1691f3ac906d4837206377765 Oct 08 00:19:53 crc kubenswrapper[4871]: I1008 00:19:53.801416 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" event={"ID":"e2f9a2a7-995e-46a7-9272-aa54c93afa60","Type":"ContainerStarted","Data":"aa8fa5332d2ef7c7d161af382a8e1c5010ed034dd832aa8910fda57a421986be"} Oct 08 00:19:53 crc kubenswrapper[4871]: I1008 00:19:53.801665 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" event={"ID":"e2f9a2a7-995e-46a7-9272-aa54c93afa60","Type":"ContainerStarted","Data":"d127f4f019be18969542bb4f9b890da62ee47ac1691f3ac906d4837206377765"} Oct 08 00:19:53 crc kubenswrapper[4871]: I1008 00:19:53.826239 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" podStartSLOduration=2.329228026 podStartE2EDuration="2.826224513s" podCreationTimestamp="2025-10-08 00:19:51 +0000 UTC" firstStartedPulling="2025-10-08 00:19:52.909177507 +0000 UTC m=+7866.711875570" lastFinishedPulling="2025-10-08 00:19:53.406173984 +0000 UTC m=+7867.208872057" observedRunningTime="2025-10-08 00:19:53.823482601 +0000 UTC m=+7867.626180684" watchObservedRunningTime="2025-10-08 00:19:53.826224513 +0000 UTC m=+7867.628922586" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.513842 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ccmpg"] Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.519515 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.531877 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ccmpg"] Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.560217 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-catalog-content\") pod \"redhat-operators-ccmpg\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.560371 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlk7s\" (UniqueName: \"kubernetes.io/projected/96868a1f-db4c-4d5f-8af7-80a149a50bc7-kube-api-access-rlk7s\") pod \"redhat-operators-ccmpg\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.560501 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-utilities\") pod \"redhat-operators-ccmpg\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.663620 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-catalog-content\") pod \"redhat-operators-ccmpg\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.663787 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlk7s\" (UniqueName: \"kubernetes.io/projected/96868a1f-db4c-4d5f-8af7-80a149a50bc7-kube-api-access-rlk7s\") pod \"redhat-operators-ccmpg\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.664015 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-utilities\") pod \"redhat-operators-ccmpg\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.664653 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-catalog-content\") pod \"redhat-operators-ccmpg\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.664673 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-utilities\") pod \"redhat-operators-ccmpg\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.693550 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlk7s\" (UniqueName: \"kubernetes.io/projected/96868a1f-db4c-4d5f-8af7-80a149a50bc7-kube-api-access-rlk7s\") pod \"redhat-operators-ccmpg\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:19:59 crc kubenswrapper[4871]: I1008 00:19:59.851145 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:20:00 crc kubenswrapper[4871]: I1008 00:20:00.448327 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ccmpg"] Oct 08 00:20:00 crc kubenswrapper[4871]: I1008 00:20:00.889723 4871 generic.go:334] "Generic (PLEG): container finished" podID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerID="b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3" exitCode=0 Oct 08 00:20:00 crc kubenswrapper[4871]: I1008 00:20:00.891063 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccmpg" event={"ID":"96868a1f-db4c-4d5f-8af7-80a149a50bc7","Type":"ContainerDied","Data":"b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3"} Oct 08 00:20:00 crc kubenswrapper[4871]: I1008 00:20:00.891105 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccmpg" event={"ID":"96868a1f-db4c-4d5f-8af7-80a149a50bc7","Type":"ContainerStarted","Data":"8109a92762f2bba896e98782ca460fb8909c0cebb27e6ce64e19f45c5fce311a"} Oct 08 00:20:02 crc kubenswrapper[4871]: I1008 00:20:02.918574 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccmpg" event={"ID":"96868a1f-db4c-4d5f-8af7-80a149a50bc7","Type":"ContainerStarted","Data":"a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a"} Oct 08 00:20:02 crc kubenswrapper[4871]: I1008 00:20:02.982616 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:20:02 crc kubenswrapper[4871]: E1008 00:20:02.982915 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:20:03 crc kubenswrapper[4871]: I1008 00:20:03.946727 4871 generic.go:334] "Generic (PLEG): container finished" podID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerID="a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a" exitCode=0 Oct 08 00:20:03 crc kubenswrapper[4871]: I1008 00:20:03.946770 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccmpg" event={"ID":"96868a1f-db4c-4d5f-8af7-80a149a50bc7","Type":"ContainerDied","Data":"a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a"} Oct 08 00:20:05 crc kubenswrapper[4871]: I1008 00:20:05.972280 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccmpg" event={"ID":"96868a1f-db4c-4d5f-8af7-80a149a50bc7","Type":"ContainerStarted","Data":"c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce"} Oct 08 00:20:05 crc kubenswrapper[4871]: I1008 00:20:05.995537 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ccmpg" podStartSLOduration=3.061646978 podStartE2EDuration="6.995516432s" podCreationTimestamp="2025-10-08 00:19:59 +0000 UTC" firstStartedPulling="2025-10-08 00:20:00.896014111 +0000 UTC m=+7874.698712184" lastFinishedPulling="2025-10-08 00:20:04.829883565 +0000 UTC m=+7878.632581638" observedRunningTime="2025-10-08 00:20:05.990041388 +0000 UTC m=+7879.792739471" watchObservedRunningTime="2025-10-08 00:20:05.995516432 +0000 UTC m=+7879.798214505" Oct 08 00:20:09 crc kubenswrapper[4871]: I1008 00:20:09.852617 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:20:09 crc kubenswrapper[4871]: I1008 00:20:09.853228 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:20:10 crc kubenswrapper[4871]: I1008 00:20:10.909245 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ccmpg" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerName="registry-server" probeResult="failure" output=< Oct 08 00:20:10 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 08 00:20:10 crc kubenswrapper[4871]: > Oct 08 00:20:15 crc kubenswrapper[4871]: I1008 00:20:15.983314 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:20:15 crc kubenswrapper[4871]: E1008 00:20:15.984166 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:20:19 crc kubenswrapper[4871]: I1008 00:20:19.930977 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:20:20 crc kubenswrapper[4871]: I1008 00:20:20.031883 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:20:20 crc kubenswrapper[4871]: I1008 00:20:20.180273 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ccmpg"] Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.161327 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ccmpg" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerName="registry-server" containerID="cri-o://c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce" gracePeriod=2 Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.787117 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.791868 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-catalog-content\") pod \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.791973 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-utilities\") pod \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.792181 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlk7s\" (UniqueName: \"kubernetes.io/projected/96868a1f-db4c-4d5f-8af7-80a149a50bc7-kube-api-access-rlk7s\") pod \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\" (UID: \"96868a1f-db4c-4d5f-8af7-80a149a50bc7\") " Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.793095 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-utilities" (OuterVolumeSpecName: "utilities") pod "96868a1f-db4c-4d5f-8af7-80a149a50bc7" (UID: "96868a1f-db4c-4d5f-8af7-80a149a50bc7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.802734 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96868a1f-db4c-4d5f-8af7-80a149a50bc7-kube-api-access-rlk7s" (OuterVolumeSpecName: "kube-api-access-rlk7s") pod "96868a1f-db4c-4d5f-8af7-80a149a50bc7" (UID: "96868a1f-db4c-4d5f-8af7-80a149a50bc7"). InnerVolumeSpecName "kube-api-access-rlk7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.897777 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.897880 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlk7s\" (UniqueName: \"kubernetes.io/projected/96868a1f-db4c-4d5f-8af7-80a149a50bc7-kube-api-access-rlk7s\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:21 crc kubenswrapper[4871]: I1008 00:20:21.927477 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96868a1f-db4c-4d5f-8af7-80a149a50bc7" (UID: "96868a1f-db4c-4d5f-8af7-80a149a50bc7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.000725 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96868a1f-db4c-4d5f-8af7-80a149a50bc7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.173666 4871 generic.go:334] "Generic (PLEG): container finished" podID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerID="c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce" exitCode=0 Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.173766 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccmpg" event={"ID":"96868a1f-db4c-4d5f-8af7-80a149a50bc7","Type":"ContainerDied","Data":"c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce"} Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.174100 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccmpg" event={"ID":"96868a1f-db4c-4d5f-8af7-80a149a50bc7","Type":"ContainerDied","Data":"8109a92762f2bba896e98782ca460fb8909c0cebb27e6ce64e19f45c5fce311a"} Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.173825 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ccmpg" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.174183 4871 scope.go:117] "RemoveContainer" containerID="c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.215750 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ccmpg"] Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.217238 4871 scope.go:117] "RemoveContainer" containerID="a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.229721 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ccmpg"] Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.252522 4871 scope.go:117] "RemoveContainer" containerID="b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.305135 4871 scope.go:117] "RemoveContainer" containerID="c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce" Oct 08 00:20:22 crc kubenswrapper[4871]: E1008 00:20:22.306601 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce\": container with ID starting with c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce not found: ID does not exist" containerID="c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.306650 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce"} err="failed to get container status \"c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce\": rpc error: code = NotFound desc = could not find container \"c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce\": container with ID starting with c766db5fd09a61be1f3bdbba6c8e816246acdf21b8d486483a73ffaa2132a3ce not found: ID does not exist" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.306679 4871 scope.go:117] "RemoveContainer" containerID="a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a" Oct 08 00:20:22 crc kubenswrapper[4871]: E1008 00:20:22.307027 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a\": container with ID starting with a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a not found: ID does not exist" containerID="a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.307055 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a"} err="failed to get container status \"a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a\": rpc error: code = NotFound desc = could not find container \"a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a\": container with ID starting with a3fea9e349a30c1897cc933cd73b1f840b495a1a9208c574072fce721d93ae0a not found: ID does not exist" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.307067 4871 scope.go:117] "RemoveContainer" containerID="b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3" Oct 08 00:20:22 crc kubenswrapper[4871]: E1008 00:20:22.307442 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3\": container with ID starting with b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3 not found: ID does not exist" containerID="b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3" Oct 08 00:20:22 crc kubenswrapper[4871]: I1008 00:20:22.307548 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3"} err="failed to get container status \"b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3\": rpc error: code = NotFound desc = could not find container \"b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3\": container with ID starting with b92adbe369c791d43f2311d3a36f942d65dc3db72c38dce57944f6b803aadda3 not found: ID does not exist" Oct 08 00:20:23 crc kubenswrapper[4871]: I1008 00:20:23.001478 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" path="/var/lib/kubelet/pods/96868a1f-db4c-4d5f-8af7-80a149a50bc7/volumes" Oct 08 00:20:29 crc kubenswrapper[4871]: I1008 00:20:29.982863 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:20:29 crc kubenswrapper[4871]: E1008 00:20:29.983755 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:20:44 crc kubenswrapper[4871]: I1008 00:20:44.983056 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:20:44 crc kubenswrapper[4871]: E1008 00:20:44.984066 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:20:49 crc kubenswrapper[4871]: I1008 00:20:49.548935 4871 generic.go:334] "Generic (PLEG): container finished" podID="e2f9a2a7-995e-46a7-9272-aa54c93afa60" containerID="aa8fa5332d2ef7c7d161af382a8e1c5010ed034dd832aa8910fda57a421986be" exitCode=0 Oct 08 00:20:49 crc kubenswrapper[4871]: I1008 00:20:49.549059 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" event={"ID":"e2f9a2a7-995e-46a7-9272-aa54c93afa60","Type":"ContainerDied","Data":"aa8fa5332d2ef7c7d161af382a8e1c5010ed034dd832aa8910fda57a421986be"} Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.136849 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.234258 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk6l6\" (UniqueName: \"kubernetes.io/projected/e2f9a2a7-995e-46a7-9272-aa54c93afa60-kube-api-access-lk6l6\") pod \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.234344 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ssh-key\") pod \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.234450 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-nova-metadata-neutron-config-0\") pod \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.234543 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-ovn-metadata-agent-neutron-config-0\") pod \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.234572 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-inventory\") pod \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.234609 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ceph\") pod \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.234646 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-metadata-combined-ca-bundle\") pod \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\" (UID: \"e2f9a2a7-995e-46a7-9272-aa54c93afa60\") " Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.240629 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e2f9a2a7-995e-46a7-9272-aa54c93afa60" (UID: "e2f9a2a7-995e-46a7-9272-aa54c93afa60"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.240691 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2f9a2a7-995e-46a7-9272-aa54c93afa60-kube-api-access-lk6l6" (OuterVolumeSpecName: "kube-api-access-lk6l6") pod "e2f9a2a7-995e-46a7-9272-aa54c93afa60" (UID: "e2f9a2a7-995e-46a7-9272-aa54c93afa60"). InnerVolumeSpecName "kube-api-access-lk6l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.243994 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ceph" (OuterVolumeSpecName: "ceph") pod "e2f9a2a7-995e-46a7-9272-aa54c93afa60" (UID: "e2f9a2a7-995e-46a7-9272-aa54c93afa60"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.265161 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e2f9a2a7-995e-46a7-9272-aa54c93afa60" (UID: "e2f9a2a7-995e-46a7-9272-aa54c93afa60"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.266847 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "e2f9a2a7-995e-46a7-9272-aa54c93afa60" (UID: "e2f9a2a7-995e-46a7-9272-aa54c93afa60"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.286636 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "e2f9a2a7-995e-46a7-9272-aa54c93afa60" (UID: "e2f9a2a7-995e-46a7-9272-aa54c93afa60"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.286769 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-inventory" (OuterVolumeSpecName: "inventory") pod "e2f9a2a7-995e-46a7-9272-aa54c93afa60" (UID: "e2f9a2a7-995e-46a7-9272-aa54c93afa60"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.337605 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk6l6\" (UniqueName: \"kubernetes.io/projected/e2f9a2a7-995e-46a7-9272-aa54c93afa60-kube-api-access-lk6l6\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.337643 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.337657 4871 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.337670 4871 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.337682 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.337693 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.337706 4871 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f9a2a7-995e-46a7-9272-aa54c93afa60-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.575100 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" event={"ID":"e2f9a2a7-995e-46a7-9272-aa54c93afa60","Type":"ContainerDied","Data":"d127f4f019be18969542bb4f9b890da62ee47ac1691f3ac906d4837206377765"} Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.575137 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d127f4f019be18969542bb4f9b890da62ee47ac1691f3ac906d4837206377765" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.575203 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-qnn9h" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.754319 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-gb8fk"] Oct 08 00:20:51 crc kubenswrapper[4871]: E1008 00:20:51.754871 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2f9a2a7-995e-46a7-9272-aa54c93afa60" containerName="neutron-metadata-openstack-openstack-cell1" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.754889 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2f9a2a7-995e-46a7-9272-aa54c93afa60" containerName="neutron-metadata-openstack-openstack-cell1" Oct 08 00:20:51 crc kubenswrapper[4871]: E1008 00:20:51.754902 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerName="extract-utilities" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.754910 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerName="extract-utilities" Oct 08 00:20:51 crc kubenswrapper[4871]: E1008 00:20:51.754937 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerName="registry-server" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.754943 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerName="registry-server" Oct 08 00:20:51 crc kubenswrapper[4871]: E1008 00:20:51.754969 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerName="extract-content" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.754975 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerName="extract-content" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.755169 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2f9a2a7-995e-46a7-9272-aa54c93afa60" containerName="neutron-metadata-openstack-openstack-cell1" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.755199 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="96868a1f-db4c-4d5f-8af7-80a149a50bc7" containerName="registry-server" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.755995 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.763049 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.763116 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.763270 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.763369 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.763530 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.775378 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-gb8fk"] Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.849998 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.850135 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ceph\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.850181 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-inventory\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.850269 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgl8j\" (UniqueName: \"kubernetes.io/projected/328f608d-5164-4165-9cfc-76a5b1caa79c-kube-api-access-vgl8j\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.850301 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.850325 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ssh-key\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.952935 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-inventory\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.953310 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgl8j\" (UniqueName: \"kubernetes.io/projected/328f608d-5164-4165-9cfc-76a5b1caa79c-kube-api-access-vgl8j\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.953344 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.953372 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ssh-key\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.953409 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.953593 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ceph\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.957709 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ssh-key\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.957763 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.957827 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ceph\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.958114 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.960024 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-inventory\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:51 crc kubenswrapper[4871]: I1008 00:20:51.972268 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgl8j\" (UniqueName: \"kubernetes.io/projected/328f608d-5164-4165-9cfc-76a5b1caa79c-kube-api-access-vgl8j\") pod \"libvirt-openstack-openstack-cell1-gb8fk\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:52 crc kubenswrapper[4871]: I1008 00:20:52.096113 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:20:52 crc kubenswrapper[4871]: I1008 00:20:52.713496 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-gb8fk"] Oct 08 00:20:52 crc kubenswrapper[4871]: W1008 00:20:52.729051 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod328f608d_5164_4165_9cfc_76a5b1caa79c.slice/crio-f7360599b841591eec4f9a97941b702fd7286821b2f482b7d617d08e9659d057 WatchSource:0}: Error finding container f7360599b841591eec4f9a97941b702fd7286821b2f482b7d617d08e9659d057: Status 404 returned error can't find the container with id f7360599b841591eec4f9a97941b702fd7286821b2f482b7d617d08e9659d057 Oct 08 00:20:53 crc kubenswrapper[4871]: I1008 00:20:53.599627 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" event={"ID":"328f608d-5164-4165-9cfc-76a5b1caa79c","Type":"ContainerStarted","Data":"70ac58919370400c684b9f8915e22df039acfdbcaaf9b3b85c1c0816c45b017c"} Oct 08 00:20:53 crc kubenswrapper[4871]: I1008 00:20:53.599964 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" event={"ID":"328f608d-5164-4165-9cfc-76a5b1caa79c","Type":"ContainerStarted","Data":"f7360599b841591eec4f9a97941b702fd7286821b2f482b7d617d08e9659d057"} Oct 08 00:20:53 crc kubenswrapper[4871]: I1008 00:20:53.625051 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" podStartSLOduration=2.181292565 podStartE2EDuration="2.625026025s" podCreationTimestamp="2025-10-08 00:20:51 +0000 UTC" firstStartedPulling="2025-10-08 00:20:52.733740614 +0000 UTC m=+7926.536438687" lastFinishedPulling="2025-10-08 00:20:53.177474084 +0000 UTC m=+7926.980172147" observedRunningTime="2025-10-08 00:20:53.617560999 +0000 UTC m=+7927.420259072" watchObservedRunningTime="2025-10-08 00:20:53.625026025 +0000 UTC m=+7927.427724098" Oct 08 00:20:59 crc kubenswrapper[4871]: I1008 00:20:59.982972 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:20:59 crc kubenswrapper[4871]: E1008 00:20:59.984081 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:21:12 crc kubenswrapper[4871]: I1008 00:21:12.982439 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:21:12 crc kubenswrapper[4871]: E1008 00:21:12.983727 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:21:23 crc kubenswrapper[4871]: I1008 00:21:23.982504 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:21:23 crc kubenswrapper[4871]: E1008 00:21:23.983693 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:21:34 crc kubenswrapper[4871]: I1008 00:21:34.982866 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:21:34 crc kubenswrapper[4871]: E1008 00:21:34.984299 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:21:48 crc kubenswrapper[4871]: I1008 00:21:48.983031 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:21:48 crc kubenswrapper[4871]: E1008 00:21:48.984019 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:22:02 crc kubenswrapper[4871]: I1008 00:22:02.982659 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:22:02 crc kubenswrapper[4871]: E1008 00:22:02.983740 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:22:14 crc kubenswrapper[4871]: I1008 00:22:14.984333 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:22:14 crc kubenswrapper[4871]: E1008 00:22:14.987283 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:22:29 crc kubenswrapper[4871]: I1008 00:22:29.984045 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:22:29 crc kubenswrapper[4871]: E1008 00:22:29.986988 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:22:41 crc kubenswrapper[4871]: I1008 00:22:41.983853 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:22:41 crc kubenswrapper[4871]: E1008 00:22:41.985665 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:22:54 crc kubenswrapper[4871]: I1008 00:22:54.983391 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:22:54 crc kubenswrapper[4871]: E1008 00:22:54.984483 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:23:08 crc kubenswrapper[4871]: I1008 00:23:08.983425 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:23:09 crc kubenswrapper[4871]: I1008 00:23:09.266148 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"41b11eaf81381092c1ab09f49f16ac15715f59b7a35ab50fd050ab813e3add3a"} Oct 08 00:23:10 crc kubenswrapper[4871]: I1008 00:23:10.941307 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c59kz"] Oct 08 00:23:10 crc kubenswrapper[4871]: I1008 00:23:10.944647 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:10 crc kubenswrapper[4871]: I1008 00:23:10.952990 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c59kz"] Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.042895 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-catalog-content\") pod \"redhat-marketplace-c59kz\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.042960 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwxqc\" (UniqueName: \"kubernetes.io/projected/76f293b7-c4a7-438f-9be7-cdb93e7f6095-kube-api-access-vwxqc\") pod \"redhat-marketplace-c59kz\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.043657 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-utilities\") pod \"redhat-marketplace-c59kz\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.145489 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-utilities\") pod \"redhat-marketplace-c59kz\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.145623 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-catalog-content\") pod \"redhat-marketplace-c59kz\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.145661 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwxqc\" (UniqueName: \"kubernetes.io/projected/76f293b7-c4a7-438f-9be7-cdb93e7f6095-kube-api-access-vwxqc\") pod \"redhat-marketplace-c59kz\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.146634 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-catalog-content\") pod \"redhat-marketplace-c59kz\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.147352 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-utilities\") pod \"redhat-marketplace-c59kz\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.172844 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwxqc\" (UniqueName: \"kubernetes.io/projected/76f293b7-c4a7-438f-9be7-cdb93e7f6095-kube-api-access-vwxqc\") pod \"redhat-marketplace-c59kz\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.276401 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:11 crc kubenswrapper[4871]: I1008 00:23:11.778090 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c59kz"] Oct 08 00:23:11 crc kubenswrapper[4871]: W1008 00:23:11.778363 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76f293b7_c4a7_438f_9be7_cdb93e7f6095.slice/crio-139f4a420ec2d80e383cde61976e738579af8bf925df41ef2371d0e0d9e8e8bb WatchSource:0}: Error finding container 139f4a420ec2d80e383cde61976e738579af8bf925df41ef2371d0e0d9e8e8bb: Status 404 returned error can't find the container with id 139f4a420ec2d80e383cde61976e738579af8bf925df41ef2371d0e0d9e8e8bb Oct 08 00:23:12 crc kubenswrapper[4871]: I1008 00:23:12.304456 4871 generic.go:334] "Generic (PLEG): container finished" podID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerID="dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366" exitCode=0 Oct 08 00:23:12 crc kubenswrapper[4871]: I1008 00:23:12.304935 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c59kz" event={"ID":"76f293b7-c4a7-438f-9be7-cdb93e7f6095","Type":"ContainerDied","Data":"dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366"} Oct 08 00:23:12 crc kubenswrapper[4871]: I1008 00:23:12.305132 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c59kz" event={"ID":"76f293b7-c4a7-438f-9be7-cdb93e7f6095","Type":"ContainerStarted","Data":"139f4a420ec2d80e383cde61976e738579af8bf925df41ef2371d0e0d9e8e8bb"} Oct 08 00:23:12 crc kubenswrapper[4871]: I1008 00:23:12.310296 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:23:13 crc kubenswrapper[4871]: I1008 00:23:13.319571 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c59kz" event={"ID":"76f293b7-c4a7-438f-9be7-cdb93e7f6095","Type":"ContainerStarted","Data":"a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40"} Oct 08 00:23:14 crc kubenswrapper[4871]: I1008 00:23:14.337346 4871 generic.go:334] "Generic (PLEG): container finished" podID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerID="a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40" exitCode=0 Oct 08 00:23:14 crc kubenswrapper[4871]: I1008 00:23:14.337665 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c59kz" event={"ID":"76f293b7-c4a7-438f-9be7-cdb93e7f6095","Type":"ContainerDied","Data":"a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40"} Oct 08 00:23:15 crc kubenswrapper[4871]: I1008 00:23:15.352887 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c59kz" event={"ID":"76f293b7-c4a7-438f-9be7-cdb93e7f6095","Type":"ContainerStarted","Data":"22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865"} Oct 08 00:23:15 crc kubenswrapper[4871]: I1008 00:23:15.385176 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c59kz" podStartSLOduration=2.890823031 podStartE2EDuration="5.385153323s" podCreationTimestamp="2025-10-08 00:23:10 +0000 UTC" firstStartedPulling="2025-10-08 00:23:12.309881892 +0000 UTC m=+8066.112579985" lastFinishedPulling="2025-10-08 00:23:14.804212174 +0000 UTC m=+8068.606910277" observedRunningTime="2025-10-08 00:23:15.377480502 +0000 UTC m=+8069.180178585" watchObservedRunningTime="2025-10-08 00:23:15.385153323 +0000 UTC m=+8069.187851416" Oct 08 00:23:21 crc kubenswrapper[4871]: I1008 00:23:21.277293 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:21 crc kubenswrapper[4871]: I1008 00:23:21.278065 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:21 crc kubenswrapper[4871]: I1008 00:23:21.360707 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:21 crc kubenswrapper[4871]: I1008 00:23:21.472684 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:21 crc kubenswrapper[4871]: I1008 00:23:21.597072 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c59kz"] Oct 08 00:23:23 crc kubenswrapper[4871]: I1008 00:23:23.438449 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c59kz" podUID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerName="registry-server" containerID="cri-o://22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865" gracePeriod=2 Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.010619 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.166705 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-catalog-content\") pod \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.167411 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-utilities\") pod \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.167765 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwxqc\" (UniqueName: \"kubernetes.io/projected/76f293b7-c4a7-438f-9be7-cdb93e7f6095-kube-api-access-vwxqc\") pod \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\" (UID: \"76f293b7-c4a7-438f-9be7-cdb93e7f6095\") " Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.168256 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-utilities" (OuterVolumeSpecName: "utilities") pod "76f293b7-c4a7-438f-9be7-cdb93e7f6095" (UID: "76f293b7-c4a7-438f-9be7-cdb93e7f6095"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.168896 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.173496 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76f293b7-c4a7-438f-9be7-cdb93e7f6095-kube-api-access-vwxqc" (OuterVolumeSpecName: "kube-api-access-vwxqc") pod "76f293b7-c4a7-438f-9be7-cdb93e7f6095" (UID: "76f293b7-c4a7-438f-9be7-cdb93e7f6095"). InnerVolumeSpecName "kube-api-access-vwxqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.179160 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76f293b7-c4a7-438f-9be7-cdb93e7f6095" (UID: "76f293b7-c4a7-438f-9be7-cdb93e7f6095"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.271265 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76f293b7-c4a7-438f-9be7-cdb93e7f6095-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.271462 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwxqc\" (UniqueName: \"kubernetes.io/projected/76f293b7-c4a7-438f-9be7-cdb93e7f6095-kube-api-access-vwxqc\") on node \"crc\" DevicePath \"\"" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.454178 4871 generic.go:334] "Generic (PLEG): container finished" podID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerID="22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865" exitCode=0 Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.454247 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c59kz" event={"ID":"76f293b7-c4a7-438f-9be7-cdb93e7f6095","Type":"ContainerDied","Data":"22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865"} Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.454582 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c59kz" event={"ID":"76f293b7-c4a7-438f-9be7-cdb93e7f6095","Type":"ContainerDied","Data":"139f4a420ec2d80e383cde61976e738579af8bf925df41ef2371d0e0d9e8e8bb"} Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.454614 4871 scope.go:117] "RemoveContainer" containerID="22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.454279 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c59kz" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.480852 4871 scope.go:117] "RemoveContainer" containerID="a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.514038 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c59kz"] Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.528838 4871 scope.go:117] "RemoveContainer" containerID="dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.529209 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c59kz"] Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.594266 4871 scope.go:117] "RemoveContainer" containerID="22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865" Oct 08 00:23:24 crc kubenswrapper[4871]: E1008 00:23:24.594895 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865\": container with ID starting with 22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865 not found: ID does not exist" containerID="22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.594995 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865"} err="failed to get container status \"22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865\": rpc error: code = NotFound desc = could not find container \"22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865\": container with ID starting with 22c7aa34530673c05a2f4151a571b4a4c7a6ead9b2d35b5ef2b3e7f712d46865 not found: ID does not exist" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.595028 4871 scope.go:117] "RemoveContainer" containerID="a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40" Oct 08 00:23:24 crc kubenswrapper[4871]: E1008 00:23:24.595425 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40\": container with ID starting with a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40 not found: ID does not exist" containerID="a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.595450 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40"} err="failed to get container status \"a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40\": rpc error: code = NotFound desc = could not find container \"a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40\": container with ID starting with a233e6e1c99e27beb2e6fc1458edde08b0d6c9c53ff9385a29d508353222ae40 not found: ID does not exist" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.595471 4871 scope.go:117] "RemoveContainer" containerID="dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366" Oct 08 00:23:24 crc kubenswrapper[4871]: E1008 00:23:24.595724 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366\": container with ID starting with dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366 not found: ID does not exist" containerID="dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366" Oct 08 00:23:24 crc kubenswrapper[4871]: I1008 00:23:24.595753 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366"} err="failed to get container status \"dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366\": rpc error: code = NotFound desc = could not find container \"dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366\": container with ID starting with dc5b8a7b545460740cedcc03719507b58ed1a0570e32a697e042337f837cc366 not found: ID does not exist" Oct 08 00:23:25 crc kubenswrapper[4871]: I1008 00:23:25.029857 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" path="/var/lib/kubelet/pods/76f293b7-c4a7-438f-9be7-cdb93e7f6095/volumes" Oct 08 00:25:35 crc kubenswrapper[4871]: I1008 00:25:35.512396 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:25:35 crc kubenswrapper[4871]: I1008 00:25:35.513058 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:25:46 crc kubenswrapper[4871]: I1008 00:25:46.123039 4871 generic.go:334] "Generic (PLEG): container finished" podID="328f608d-5164-4165-9cfc-76a5b1caa79c" containerID="70ac58919370400c684b9f8915e22df039acfdbcaaf9b3b85c1c0816c45b017c" exitCode=0 Oct 08 00:25:46 crc kubenswrapper[4871]: I1008 00:25:46.123132 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" event={"ID":"328f608d-5164-4165-9cfc-76a5b1caa79c","Type":"ContainerDied","Data":"70ac58919370400c684b9f8915e22df039acfdbcaaf9b3b85c1c0816c45b017c"} Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.680417 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.833455 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-secret-0\") pod \"328f608d-5164-4165-9cfc-76a5b1caa79c\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.833979 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-combined-ca-bundle\") pod \"328f608d-5164-4165-9cfc-76a5b1caa79c\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.834058 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-inventory\") pod \"328f608d-5164-4165-9cfc-76a5b1caa79c\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.834082 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ssh-key\") pod \"328f608d-5164-4165-9cfc-76a5b1caa79c\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.834137 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgl8j\" (UniqueName: \"kubernetes.io/projected/328f608d-5164-4165-9cfc-76a5b1caa79c-kube-api-access-vgl8j\") pod \"328f608d-5164-4165-9cfc-76a5b1caa79c\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.834267 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ceph\") pod \"328f608d-5164-4165-9cfc-76a5b1caa79c\" (UID: \"328f608d-5164-4165-9cfc-76a5b1caa79c\") " Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.841243 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ceph" (OuterVolumeSpecName: "ceph") pod "328f608d-5164-4165-9cfc-76a5b1caa79c" (UID: "328f608d-5164-4165-9cfc-76a5b1caa79c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.841444 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/328f608d-5164-4165-9cfc-76a5b1caa79c-kube-api-access-vgl8j" (OuterVolumeSpecName: "kube-api-access-vgl8j") pod "328f608d-5164-4165-9cfc-76a5b1caa79c" (UID: "328f608d-5164-4165-9cfc-76a5b1caa79c"). InnerVolumeSpecName "kube-api-access-vgl8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.841737 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "328f608d-5164-4165-9cfc-76a5b1caa79c" (UID: "328f608d-5164-4165-9cfc-76a5b1caa79c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.867988 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "328f608d-5164-4165-9cfc-76a5b1caa79c" (UID: "328f608d-5164-4165-9cfc-76a5b1caa79c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.887119 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-inventory" (OuterVolumeSpecName: "inventory") pod "328f608d-5164-4165-9cfc-76a5b1caa79c" (UID: "328f608d-5164-4165-9cfc-76a5b1caa79c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.889260 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "328f608d-5164-4165-9cfc-76a5b1caa79c" (UID: "328f608d-5164-4165-9cfc-76a5b1caa79c"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.936699 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.936740 4871 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.936754 4871 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.936768 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.936781 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328f608d-5164-4165-9cfc-76a5b1caa79c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:47 crc kubenswrapper[4871]: I1008 00:25:47.936808 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgl8j\" (UniqueName: \"kubernetes.io/projected/328f608d-5164-4165-9cfc-76a5b1caa79c-kube-api-access-vgl8j\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.163318 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.163356 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-gb8fk" event={"ID":"328f608d-5164-4165-9cfc-76a5b1caa79c","Type":"ContainerDied","Data":"f7360599b841591eec4f9a97941b702fd7286821b2f482b7d617d08e9659d057"} Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.163496 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7360599b841591eec4f9a97941b702fd7286821b2f482b7d617d08e9659d057" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.242002 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-k2kcq"] Oct 08 00:25:48 crc kubenswrapper[4871]: E1008 00:25:48.242715 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="328f608d-5164-4165-9cfc-76a5b1caa79c" containerName="libvirt-openstack-openstack-cell1" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.242848 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="328f608d-5164-4165-9cfc-76a5b1caa79c" containerName="libvirt-openstack-openstack-cell1" Oct 08 00:25:48 crc kubenswrapper[4871]: E1008 00:25:48.242969 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerName="extract-utilities" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.243051 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerName="extract-utilities" Oct 08 00:25:48 crc kubenswrapper[4871]: E1008 00:25:48.243139 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerName="registry-server" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.243207 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerName="registry-server" Oct 08 00:25:48 crc kubenswrapper[4871]: E1008 00:25:48.243298 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerName="extract-content" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.243365 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerName="extract-content" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.243712 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="328f608d-5164-4165-9cfc-76a5b1caa79c" containerName="libvirt-openstack-openstack-cell1" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.243866 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f293b7-c4a7-438f-9be7-cdb93e7f6095" containerName="registry-server" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.244870 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.247145 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.247842 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.248056 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.248147 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.250539 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.251353 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.252523 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.278295 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-k2kcq"] Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.349131 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.349263 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ceph\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.349295 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.349715 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.349835 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f996c\" (UniqueName: \"kubernetes.io/projected/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-kube-api-access-f996c\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.349881 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.349908 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.349927 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-inventory\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.349947 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.350009 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.350036 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451533 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451591 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f996c\" (UniqueName: \"kubernetes.io/projected/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-kube-api-access-f996c\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451624 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451651 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451678 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451703 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-inventory\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451767 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451825 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451883 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451944 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ceph\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.451982 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.453226 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.453246 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.455842 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.456743 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ceph\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.458944 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-inventory\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.459926 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.460307 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.469100 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.470044 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.472576 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.479378 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f996c\" (UniqueName: \"kubernetes.io/projected/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-kube-api-access-f996c\") pod \"nova-cell1-openstack-openstack-cell1-k2kcq\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:48 crc kubenswrapper[4871]: I1008 00:25:48.574356 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:25:49 crc kubenswrapper[4871]: I1008 00:25:49.161833 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-k2kcq"] Oct 08 00:25:49 crc kubenswrapper[4871]: I1008 00:25:49.173001 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" event={"ID":"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe","Type":"ContainerStarted","Data":"f216a0320bb08f8e880496998332e97742e419e0e82540347235641a6f1a230b"} Oct 08 00:25:50 crc kubenswrapper[4871]: I1008 00:25:50.186973 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" event={"ID":"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe","Type":"ContainerStarted","Data":"b76467911db8c22ffb51cdadf58be26bdb9744671c9aa7c84b83eb16454a89e5"} Oct 08 00:25:50 crc kubenswrapper[4871]: I1008 00:25:50.212733 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" podStartSLOduration=1.628656211 podStartE2EDuration="2.212685438s" podCreationTimestamp="2025-10-08 00:25:48 +0000 UTC" firstStartedPulling="2025-10-08 00:25:49.165411748 +0000 UTC m=+8222.968109821" lastFinishedPulling="2025-10-08 00:25:49.749440955 +0000 UTC m=+8223.552139048" observedRunningTime="2025-10-08 00:25:50.206943538 +0000 UTC m=+8224.009641631" watchObservedRunningTime="2025-10-08 00:25:50.212685438 +0000 UTC m=+8224.015383511" Oct 08 00:26:05 crc kubenswrapper[4871]: I1008 00:26:05.512822 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:26:05 crc kubenswrapper[4871]: I1008 00:26:05.513602 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:26:35 crc kubenswrapper[4871]: I1008 00:26:35.512399 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:26:35 crc kubenswrapper[4871]: I1008 00:26:35.513146 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:26:35 crc kubenswrapper[4871]: I1008 00:26:35.513229 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:26:35 crc kubenswrapper[4871]: I1008 00:26:35.514826 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41b11eaf81381092c1ab09f49f16ac15715f59b7a35ab50fd050ab813e3add3a"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:26:35 crc kubenswrapper[4871]: I1008 00:26:35.514957 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://41b11eaf81381092c1ab09f49f16ac15715f59b7a35ab50fd050ab813e3add3a" gracePeriod=600 Oct 08 00:26:35 crc kubenswrapper[4871]: I1008 00:26:35.699937 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="41b11eaf81381092c1ab09f49f16ac15715f59b7a35ab50fd050ab813e3add3a" exitCode=0 Oct 08 00:26:35 crc kubenswrapper[4871]: I1008 00:26:35.700046 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"41b11eaf81381092c1ab09f49f16ac15715f59b7a35ab50fd050ab813e3add3a"} Oct 08 00:26:35 crc kubenswrapper[4871]: I1008 00:26:35.700363 4871 scope.go:117] "RemoveContainer" containerID="2d33b2c989adc42a06a7d8b1f696d62a47ada72ea21afec85fc0c90e7bf023ed" Oct 08 00:26:36 crc kubenswrapper[4871]: I1008 00:26:36.716181 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d"} Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.433241 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mvvfq"] Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.436001 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.446945 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mvvfq"] Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.514161 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-utilities\") pod \"certified-operators-mvvfq\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.514226 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-catalog-content\") pod \"certified-operators-mvvfq\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.514256 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8xhq\" (UniqueName: \"kubernetes.io/projected/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-kube-api-access-v8xhq\") pod \"certified-operators-mvvfq\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.616448 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-utilities\") pod \"certified-operators-mvvfq\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.616506 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-catalog-content\") pod \"certified-operators-mvvfq\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.616544 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8xhq\" (UniqueName: \"kubernetes.io/projected/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-kube-api-access-v8xhq\") pod \"certified-operators-mvvfq\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.616941 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-utilities\") pod \"certified-operators-mvvfq\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.617091 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-catalog-content\") pod \"certified-operators-mvvfq\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.636584 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8xhq\" (UniqueName: \"kubernetes.io/projected/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-kube-api-access-v8xhq\") pod \"certified-operators-mvvfq\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:03 crc kubenswrapper[4871]: I1008 00:27:03.756948 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:04 crc kubenswrapper[4871]: I1008 00:27:04.290327 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mvvfq"] Oct 08 00:27:05 crc kubenswrapper[4871]: I1008 00:27:05.053865 4871 generic.go:334] "Generic (PLEG): container finished" podID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerID="41a0c0261fee64744ddb7c9770d6aa8b8b83c9a2b7ccb8ee9c0836349df13a97" exitCode=0 Oct 08 00:27:05 crc kubenswrapper[4871]: I1008 00:27:05.053973 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvvfq" event={"ID":"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb","Type":"ContainerDied","Data":"41a0c0261fee64744ddb7c9770d6aa8b8b83c9a2b7ccb8ee9c0836349df13a97"} Oct 08 00:27:05 crc kubenswrapper[4871]: I1008 00:27:05.054197 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvvfq" event={"ID":"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb","Type":"ContainerStarted","Data":"cacb202e982f114ff64a5b933cf92d9feed64f9f47c10263830fd86bd79b523b"} Oct 08 00:27:06 crc kubenswrapper[4871]: I1008 00:27:06.070532 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvvfq" event={"ID":"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb","Type":"ContainerStarted","Data":"60b9876d18da8bc0c27df4c9ae0bf086882e6f7a48b7974621fd16be4af23fd0"} Oct 08 00:27:07 crc kubenswrapper[4871]: I1008 00:27:07.083203 4871 generic.go:334] "Generic (PLEG): container finished" podID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerID="60b9876d18da8bc0c27df4c9ae0bf086882e6f7a48b7974621fd16be4af23fd0" exitCode=0 Oct 08 00:27:07 crc kubenswrapper[4871]: I1008 00:27:07.083325 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvvfq" event={"ID":"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb","Type":"ContainerDied","Data":"60b9876d18da8bc0c27df4c9ae0bf086882e6f7a48b7974621fd16be4af23fd0"} Oct 08 00:27:08 crc kubenswrapper[4871]: I1008 00:27:08.099216 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvvfq" event={"ID":"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb","Type":"ContainerStarted","Data":"95979ff67ff53abf10499f100d18cf45c6e20cf7ee01fddaf514ad6698accd36"} Oct 08 00:27:08 crc kubenswrapper[4871]: I1008 00:27:08.129528 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mvvfq" podStartSLOduration=2.398238162 podStartE2EDuration="5.129503853s" podCreationTimestamp="2025-10-08 00:27:03 +0000 UTC" firstStartedPulling="2025-10-08 00:27:05.056427159 +0000 UTC m=+8298.859125232" lastFinishedPulling="2025-10-08 00:27:07.78769286 +0000 UTC m=+8301.590390923" observedRunningTime="2025-10-08 00:27:08.119020259 +0000 UTC m=+8301.921718342" watchObservedRunningTime="2025-10-08 00:27:08.129503853 +0000 UTC m=+8301.932201936" Oct 08 00:27:13 crc kubenswrapper[4871]: I1008 00:27:13.757947 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:13 crc kubenswrapper[4871]: I1008 00:27:13.758737 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:13 crc kubenswrapper[4871]: I1008 00:27:13.844620 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:14 crc kubenswrapper[4871]: I1008 00:27:14.252944 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:14 crc kubenswrapper[4871]: I1008 00:27:14.305448 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mvvfq"] Oct 08 00:27:16 crc kubenswrapper[4871]: I1008 00:27:16.198225 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mvvfq" podUID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerName="registry-server" containerID="cri-o://95979ff67ff53abf10499f100d18cf45c6e20cf7ee01fddaf514ad6698accd36" gracePeriod=2 Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.214428 4871 generic.go:334] "Generic (PLEG): container finished" podID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerID="95979ff67ff53abf10499f100d18cf45c6e20cf7ee01fddaf514ad6698accd36" exitCode=0 Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.214517 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvvfq" event={"ID":"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb","Type":"ContainerDied","Data":"95979ff67ff53abf10499f100d18cf45c6e20cf7ee01fddaf514ad6698accd36"} Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.215332 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvvfq" event={"ID":"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb","Type":"ContainerDied","Data":"cacb202e982f114ff64a5b933cf92d9feed64f9f47c10263830fd86bd79b523b"} Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.215360 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cacb202e982f114ff64a5b933cf92d9feed64f9f47c10263830fd86bd79b523b" Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.239717 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.334389 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-utilities\") pod \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.334489 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-catalog-content\") pod \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.334753 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8xhq\" (UniqueName: \"kubernetes.io/projected/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-kube-api-access-v8xhq\") pod \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\" (UID: \"04fc5299-5bd9-4d6a-b6ba-05afb34af7fb\") " Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.335427 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-utilities" (OuterVolumeSpecName: "utilities") pod "04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" (UID: "04fc5299-5bd9-4d6a-b6ba-05afb34af7fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.335628 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.342197 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-kube-api-access-v8xhq" (OuterVolumeSpecName: "kube-api-access-v8xhq") pod "04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" (UID: "04fc5299-5bd9-4d6a-b6ba-05afb34af7fb"). InnerVolumeSpecName "kube-api-access-v8xhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.385239 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" (UID: "04fc5299-5bd9-4d6a-b6ba-05afb34af7fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.438371 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:27:17 crc kubenswrapper[4871]: I1008 00:27:17.438835 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8xhq\" (UniqueName: \"kubernetes.io/projected/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb-kube-api-access-v8xhq\") on node \"crc\" DevicePath \"\"" Oct 08 00:27:18 crc kubenswrapper[4871]: I1008 00:27:18.226396 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvvfq" Oct 08 00:27:18 crc kubenswrapper[4871]: I1008 00:27:18.293576 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mvvfq"] Oct 08 00:27:18 crc kubenswrapper[4871]: I1008 00:27:18.308099 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mvvfq"] Oct 08 00:27:19 crc kubenswrapper[4871]: I1008 00:27:19.003723 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" path="/var/lib/kubelet/pods/04fc5299-5bd9-4d6a-b6ba-05afb34af7fb/volumes" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.724610 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wq48w"] Oct 08 00:27:20 crc kubenswrapper[4871]: E1008 00:27:20.725273 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerName="registry-server" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.725294 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerName="registry-server" Oct 08 00:27:20 crc kubenswrapper[4871]: E1008 00:27:20.725314 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerName="extract-content" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.725321 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerName="extract-content" Oct 08 00:27:20 crc kubenswrapper[4871]: E1008 00:27:20.725355 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerName="extract-utilities" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.725381 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerName="extract-utilities" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.725634 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="04fc5299-5bd9-4d6a-b6ba-05afb34af7fb" containerName="registry-server" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.727725 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.752637 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wq48w"] Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.810310 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-catalog-content\") pod \"community-operators-wq48w\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.810440 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwq4q\" (UniqueName: \"kubernetes.io/projected/658086f4-225a-4086-8dab-7af02ca7ad8d-kube-api-access-pwq4q\") pod \"community-operators-wq48w\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.810881 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-utilities\") pod \"community-operators-wq48w\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.913022 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwq4q\" (UniqueName: \"kubernetes.io/projected/658086f4-225a-4086-8dab-7af02ca7ad8d-kube-api-access-pwq4q\") pod \"community-operators-wq48w\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.913200 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-utilities\") pod \"community-operators-wq48w\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.913306 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-catalog-content\") pod \"community-operators-wq48w\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.913808 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-utilities\") pod \"community-operators-wq48w\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.913882 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-catalog-content\") pod \"community-operators-wq48w\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:20 crc kubenswrapper[4871]: I1008 00:27:20.946522 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwq4q\" (UniqueName: \"kubernetes.io/projected/658086f4-225a-4086-8dab-7af02ca7ad8d-kube-api-access-pwq4q\") pod \"community-operators-wq48w\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:21 crc kubenswrapper[4871]: I1008 00:27:21.066563 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:21 crc kubenswrapper[4871]: W1008 00:27:21.672070 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod658086f4_225a_4086_8dab_7af02ca7ad8d.slice/crio-982d9cd8ec8c41567a48b20c360ce221dadd4f797be14e365d7911e8906311b5 WatchSource:0}: Error finding container 982d9cd8ec8c41567a48b20c360ce221dadd4f797be14e365d7911e8906311b5: Status 404 returned error can't find the container with id 982d9cd8ec8c41567a48b20c360ce221dadd4f797be14e365d7911e8906311b5 Oct 08 00:27:21 crc kubenswrapper[4871]: I1008 00:27:21.673065 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wq48w"] Oct 08 00:27:22 crc kubenswrapper[4871]: I1008 00:27:22.294016 4871 generic.go:334] "Generic (PLEG): container finished" podID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerID="562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740" exitCode=0 Oct 08 00:27:22 crc kubenswrapper[4871]: I1008 00:27:22.294106 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wq48w" event={"ID":"658086f4-225a-4086-8dab-7af02ca7ad8d","Type":"ContainerDied","Data":"562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740"} Oct 08 00:27:22 crc kubenswrapper[4871]: I1008 00:27:22.294381 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wq48w" event={"ID":"658086f4-225a-4086-8dab-7af02ca7ad8d","Type":"ContainerStarted","Data":"982d9cd8ec8c41567a48b20c360ce221dadd4f797be14e365d7911e8906311b5"} Oct 08 00:27:23 crc kubenswrapper[4871]: I1008 00:27:23.309165 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wq48w" event={"ID":"658086f4-225a-4086-8dab-7af02ca7ad8d","Type":"ContainerStarted","Data":"a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839"} Oct 08 00:27:24 crc kubenswrapper[4871]: I1008 00:27:24.322137 4871 generic.go:334] "Generic (PLEG): container finished" podID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerID="a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839" exitCode=0 Oct 08 00:27:24 crc kubenswrapper[4871]: I1008 00:27:24.322195 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wq48w" event={"ID":"658086f4-225a-4086-8dab-7af02ca7ad8d","Type":"ContainerDied","Data":"a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839"} Oct 08 00:27:25 crc kubenswrapper[4871]: I1008 00:27:25.333365 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wq48w" event={"ID":"658086f4-225a-4086-8dab-7af02ca7ad8d","Type":"ContainerStarted","Data":"3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04"} Oct 08 00:27:25 crc kubenswrapper[4871]: I1008 00:27:25.358341 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wq48w" podStartSLOduration=2.910980299 podStartE2EDuration="5.358321009s" podCreationTimestamp="2025-10-08 00:27:20 +0000 UTC" firstStartedPulling="2025-10-08 00:27:22.298033899 +0000 UTC m=+8316.100732002" lastFinishedPulling="2025-10-08 00:27:24.745374639 +0000 UTC m=+8318.548072712" observedRunningTime="2025-10-08 00:27:25.352159958 +0000 UTC m=+8319.154858021" watchObservedRunningTime="2025-10-08 00:27:25.358321009 +0000 UTC m=+8319.161019082" Oct 08 00:27:31 crc kubenswrapper[4871]: I1008 00:27:31.067113 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:31 crc kubenswrapper[4871]: I1008 00:27:31.068179 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:31 crc kubenswrapper[4871]: I1008 00:27:31.138098 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:31 crc kubenswrapper[4871]: I1008 00:27:31.455261 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:31 crc kubenswrapper[4871]: I1008 00:27:31.515826 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wq48w"] Oct 08 00:27:33 crc kubenswrapper[4871]: I1008 00:27:33.418720 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wq48w" podUID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerName="registry-server" containerID="cri-o://3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04" gracePeriod=2 Oct 08 00:27:33 crc kubenswrapper[4871]: I1008 00:27:33.917074 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.053733 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-utilities\") pod \"658086f4-225a-4086-8dab-7af02ca7ad8d\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.054220 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-catalog-content\") pod \"658086f4-225a-4086-8dab-7af02ca7ad8d\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.055057 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwq4q\" (UniqueName: \"kubernetes.io/projected/658086f4-225a-4086-8dab-7af02ca7ad8d-kube-api-access-pwq4q\") pod \"658086f4-225a-4086-8dab-7af02ca7ad8d\" (UID: \"658086f4-225a-4086-8dab-7af02ca7ad8d\") " Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.056125 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-utilities" (OuterVolumeSpecName: "utilities") pod "658086f4-225a-4086-8dab-7af02ca7ad8d" (UID: "658086f4-225a-4086-8dab-7af02ca7ad8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.056579 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.061941 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/658086f4-225a-4086-8dab-7af02ca7ad8d-kube-api-access-pwq4q" (OuterVolumeSpecName: "kube-api-access-pwq4q") pod "658086f4-225a-4086-8dab-7af02ca7ad8d" (UID: "658086f4-225a-4086-8dab-7af02ca7ad8d"). InnerVolumeSpecName "kube-api-access-pwq4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.123511 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "658086f4-225a-4086-8dab-7af02ca7ad8d" (UID: "658086f4-225a-4086-8dab-7af02ca7ad8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.158509 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658086f4-225a-4086-8dab-7af02ca7ad8d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.158538 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwq4q\" (UniqueName: \"kubernetes.io/projected/658086f4-225a-4086-8dab-7af02ca7ad8d-kube-api-access-pwq4q\") on node \"crc\" DevicePath \"\"" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.431018 4871 generic.go:334] "Generic (PLEG): container finished" podID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerID="3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04" exitCode=0 Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.431062 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wq48w" event={"ID":"658086f4-225a-4086-8dab-7af02ca7ad8d","Type":"ContainerDied","Data":"3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04"} Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.431096 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wq48w" event={"ID":"658086f4-225a-4086-8dab-7af02ca7ad8d","Type":"ContainerDied","Data":"982d9cd8ec8c41567a48b20c360ce221dadd4f797be14e365d7911e8906311b5"} Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.431117 4871 scope.go:117] "RemoveContainer" containerID="3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.432059 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wq48w" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.474432 4871 scope.go:117] "RemoveContainer" containerID="a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.483050 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wq48w"] Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.491493 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wq48w"] Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.502142 4871 scope.go:117] "RemoveContainer" containerID="562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.546825 4871 scope.go:117] "RemoveContainer" containerID="3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04" Oct 08 00:27:34 crc kubenswrapper[4871]: E1008 00:27:34.547410 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04\": container with ID starting with 3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04 not found: ID does not exist" containerID="3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.547500 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04"} err="failed to get container status \"3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04\": rpc error: code = NotFound desc = could not find container \"3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04\": container with ID starting with 3de055de95c47ef99ae4437f9fce2a65c9309f87fa2b084544e3275a7d6c7f04 not found: ID does not exist" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.547558 4871 scope.go:117] "RemoveContainer" containerID="a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839" Oct 08 00:27:34 crc kubenswrapper[4871]: E1008 00:27:34.548166 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839\": container with ID starting with a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839 not found: ID does not exist" containerID="a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.548218 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839"} err="failed to get container status \"a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839\": rpc error: code = NotFound desc = could not find container \"a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839\": container with ID starting with a612f4f0fca061f510068606886fa7cbd3ace0785fc37a46788835f34d142839 not found: ID does not exist" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.548251 4871 scope.go:117] "RemoveContainer" containerID="562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740" Oct 08 00:27:34 crc kubenswrapper[4871]: E1008 00:27:34.548709 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740\": container with ID starting with 562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740 not found: ID does not exist" containerID="562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740" Oct 08 00:27:34 crc kubenswrapper[4871]: I1008 00:27:34.548822 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740"} err="failed to get container status \"562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740\": rpc error: code = NotFound desc = could not find container \"562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740\": container with ID starting with 562452129ba9fb2d2f2c1fd7ac95a76b843989f9aa19b51181fc7a8d771c2740 not found: ID does not exist" Oct 08 00:27:35 crc kubenswrapper[4871]: I1008 00:27:35.001676 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="658086f4-225a-4086-8dab-7af02ca7ad8d" path="/var/lib/kubelet/pods/658086f4-225a-4086-8dab-7af02ca7ad8d/volumes" Oct 08 00:28:35 crc kubenswrapper[4871]: I1008 00:28:35.513166 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:28:35 crc kubenswrapper[4871]: I1008 00:28:35.515955 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:29:05 crc kubenswrapper[4871]: I1008 00:29:05.512360 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:29:05 crc kubenswrapper[4871]: I1008 00:29:05.513152 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:29:35 crc kubenswrapper[4871]: I1008 00:29:35.512359 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:29:35 crc kubenswrapper[4871]: I1008 00:29:35.513089 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:29:35 crc kubenswrapper[4871]: I1008 00:29:35.513155 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:29:35 crc kubenswrapper[4871]: I1008 00:29:35.514377 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:29:35 crc kubenswrapper[4871]: I1008 00:29:35.514478 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" gracePeriod=600 Oct 08 00:29:35 crc kubenswrapper[4871]: E1008 00:29:35.642454 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:29:35 crc kubenswrapper[4871]: I1008 00:29:35.932376 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" exitCode=0 Oct 08 00:29:35 crc kubenswrapper[4871]: I1008 00:29:35.932447 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d"} Oct 08 00:29:35 crc kubenswrapper[4871]: I1008 00:29:35.932926 4871 scope.go:117] "RemoveContainer" containerID="41b11eaf81381092c1ab09f49f16ac15715f59b7a35ab50fd050ab813e3add3a" Oct 08 00:29:35 crc kubenswrapper[4871]: I1008 00:29:35.933653 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:29:35 crc kubenswrapper[4871]: E1008 00:29:35.933935 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:29:41 crc kubenswrapper[4871]: I1008 00:29:41.039458 4871 generic.go:334] "Generic (PLEG): container finished" podID="bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" containerID="b76467911db8c22ffb51cdadf58be26bdb9744671c9aa7c84b83eb16454a89e5" exitCode=0 Oct 08 00:29:41 crc kubenswrapper[4871]: I1008 00:29:41.039566 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" event={"ID":"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe","Type":"ContainerDied","Data":"b76467911db8c22ffb51cdadf58be26bdb9744671c9aa7c84b83eb16454a89e5"} Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.516551 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631070 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-0\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631124 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-1\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631175 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-0\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631210 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ceph\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631229 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-1\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631267 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-combined-ca-bundle\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631285 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-1\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631301 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f996c\" (UniqueName: \"kubernetes.io/projected/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-kube-api-access-f996c\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631525 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-0\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631604 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ssh-key\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.631667 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-inventory\") pod \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\" (UID: \"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe\") " Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.657876 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-kube-api-access-f996c" (OuterVolumeSpecName: "kube-api-access-f996c") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "kube-api-access-f996c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.665385 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ceph" (OuterVolumeSpecName: "ceph") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.672526 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.679515 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.683860 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.693141 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.694494 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-inventory" (OuterVolumeSpecName: "inventory") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.695579 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.699234 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.702480 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.706976 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" (UID: "bf1b098f-7fd8-4c0d-8c4e-88abab628cbe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734449 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734478 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734487 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734495 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734503 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734513 4871 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734521 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734529 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734538 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734546 4871 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:42 crc kubenswrapper[4871]: I1008 00:29:42.734555 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f996c\" (UniqueName: \"kubernetes.io/projected/bf1b098f-7fd8-4c0d-8c4e-88abab628cbe-kube-api-access-f996c\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.066373 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" event={"ID":"bf1b098f-7fd8-4c0d-8c4e-88abab628cbe","Type":"ContainerDied","Data":"f216a0320bb08f8e880496998332e97742e419e0e82540347235641a6f1a230b"} Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.066655 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f216a0320bb08f8e880496998332e97742e419e0e82540347235641a6f1a230b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.066501 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-k2kcq" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.239712 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-jmt5b"] Oct 08 00:29:43 crc kubenswrapper[4871]: E1008 00:29:43.240384 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerName="extract-content" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.240409 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerName="extract-content" Oct 08 00:29:43 crc kubenswrapper[4871]: E1008 00:29:43.240452 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" containerName="nova-cell1-openstack-openstack-cell1" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.240462 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" containerName="nova-cell1-openstack-openstack-cell1" Oct 08 00:29:43 crc kubenswrapper[4871]: E1008 00:29:43.240480 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerName="extract-utilities" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.240488 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerName="extract-utilities" Oct 08 00:29:43 crc kubenswrapper[4871]: E1008 00:29:43.240525 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerName="registry-server" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.240533 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerName="registry-server" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.251199 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf1b098f-7fd8-4c0d-8c4e-88abab628cbe" containerName="nova-cell1-openstack-openstack-cell1" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.251246 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="658086f4-225a-4086-8dab-7af02ca7ad8d" containerName="registry-server" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.252427 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.255982 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.256499 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.256807 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.257052 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.257267 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.278648 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-jmt5b"] Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.350002 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p2l8\" (UniqueName: \"kubernetes.io/projected/7b13172c-c2ed-41f2-ab24-538418c2e1b3-kube-api-access-9p2l8\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.350065 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceph\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.350097 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.350125 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.350147 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-inventory\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.350164 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ssh-key\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.350230 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.350307 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.451886 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p2l8\" (UniqueName: \"kubernetes.io/projected/7b13172c-c2ed-41f2-ab24-538418c2e1b3-kube-api-access-9p2l8\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.451952 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceph\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.451983 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.452010 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.452032 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-inventory\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.452052 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ssh-key\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.452117 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.452192 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.458192 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.465371 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.465499 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-inventory\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.465832 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.466126 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceph\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.466165 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ssh-key\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.467345 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.482027 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p2l8\" (UniqueName: \"kubernetes.io/projected/7b13172c-c2ed-41f2-ab24-538418c2e1b3-kube-api-access-9p2l8\") pod \"telemetry-openstack-openstack-cell1-jmt5b\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:43 crc kubenswrapper[4871]: I1008 00:29:43.576508 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:29:44 crc kubenswrapper[4871]: I1008 00:29:44.135025 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-jmt5b"] Oct 08 00:29:44 crc kubenswrapper[4871]: I1008 00:29:44.147335 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:29:45 crc kubenswrapper[4871]: I1008 00:29:45.105814 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" event={"ID":"7b13172c-c2ed-41f2-ab24-538418c2e1b3","Type":"ContainerStarted","Data":"4954327643ffea2cfe7633b99dd5682b27f2d5c4115c22069ad0b0dfcc0b3e21"} Oct 08 00:29:46 crc kubenswrapper[4871]: I1008 00:29:46.120488 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" event={"ID":"7b13172c-c2ed-41f2-ab24-538418c2e1b3","Type":"ContainerStarted","Data":"24803e34d3a8239446f1a413315e33e5bd967f2674f4d78e9281bc969e0d9e21"} Oct 08 00:29:46 crc kubenswrapper[4871]: I1008 00:29:46.150391 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" podStartSLOduration=2.320732205 podStartE2EDuration="3.150366852s" podCreationTimestamp="2025-10-08 00:29:43 +0000 UTC" firstStartedPulling="2025-10-08 00:29:44.147049245 +0000 UTC m=+8457.949747328" lastFinishedPulling="2025-10-08 00:29:44.976683902 +0000 UTC m=+8458.779381975" observedRunningTime="2025-10-08 00:29:46.147082706 +0000 UTC m=+8459.949780799" watchObservedRunningTime="2025-10-08 00:29:46.150366852 +0000 UTC m=+8459.953064965" Oct 08 00:29:49 crc kubenswrapper[4871]: I1008 00:29:49.982717 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:29:49 crc kubenswrapper[4871]: E1008 00:29:49.984711 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.164990 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh"] Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.169443 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.173259 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.173453 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.193248 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh"] Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.345207 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjzdn\" (UniqueName: \"kubernetes.io/projected/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-kube-api-access-bjzdn\") pod \"collect-profiles-29331390-zt7bh\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.345311 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-config-volume\") pod \"collect-profiles-29331390-zt7bh\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.345352 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-secret-volume\") pod \"collect-profiles-29331390-zt7bh\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.447287 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjzdn\" (UniqueName: \"kubernetes.io/projected/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-kube-api-access-bjzdn\") pod \"collect-profiles-29331390-zt7bh\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.447651 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-config-volume\") pod \"collect-profiles-29331390-zt7bh\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.447811 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-secret-volume\") pod \"collect-profiles-29331390-zt7bh\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.448941 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-config-volume\") pod \"collect-profiles-29331390-zt7bh\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.463810 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-secret-volume\") pod \"collect-profiles-29331390-zt7bh\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.477178 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjzdn\" (UniqueName: \"kubernetes.io/projected/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-kube-api-access-bjzdn\") pod \"collect-profiles-29331390-zt7bh\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:00 crc kubenswrapper[4871]: I1008 00:30:00.504730 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:01 crc kubenswrapper[4871]: I1008 00:30:01.032071 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh"] Oct 08 00:30:01 crc kubenswrapper[4871]: I1008 00:30:01.308060 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" event={"ID":"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be","Type":"ContainerStarted","Data":"49e4866bd7ca593719aca703247dfd64cbd2fe1e7d8b742d270157760d29aff2"} Oct 08 00:30:02 crc kubenswrapper[4871]: I1008 00:30:02.322039 4871 generic.go:334] "Generic (PLEG): container finished" podID="e601ded9-3c4a-4ddf-b5a1-c27eb017e8be" containerID="c725ec03202368799656de5b90ca733bde9f16f55a1c1148ba3915346f00fda1" exitCode=0 Oct 08 00:30:02 crc kubenswrapper[4871]: I1008 00:30:02.322366 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" event={"ID":"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be","Type":"ContainerDied","Data":"c725ec03202368799656de5b90ca733bde9f16f55a1c1148ba3915346f00fda1"} Oct 08 00:30:03 crc kubenswrapper[4871]: I1008 00:30:03.770591 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:03 crc kubenswrapper[4871]: I1008 00:30:03.926696 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjzdn\" (UniqueName: \"kubernetes.io/projected/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-kube-api-access-bjzdn\") pod \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " Oct 08 00:30:03 crc kubenswrapper[4871]: I1008 00:30:03.926810 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-config-volume\") pod \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " Oct 08 00:30:03 crc kubenswrapper[4871]: I1008 00:30:03.926998 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-secret-volume\") pod \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\" (UID: \"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be\") " Oct 08 00:30:03 crc kubenswrapper[4871]: I1008 00:30:03.927628 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-config-volume" (OuterVolumeSpecName: "config-volume") pod "e601ded9-3c4a-4ddf-b5a1-c27eb017e8be" (UID: "e601ded9-3c4a-4ddf-b5a1-c27eb017e8be"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:30:03 crc kubenswrapper[4871]: I1008 00:30:03.932679 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-kube-api-access-bjzdn" (OuterVolumeSpecName: "kube-api-access-bjzdn") pod "e601ded9-3c4a-4ddf-b5a1-c27eb017e8be" (UID: "e601ded9-3c4a-4ddf-b5a1-c27eb017e8be"). InnerVolumeSpecName "kube-api-access-bjzdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:30:03 crc kubenswrapper[4871]: I1008 00:30:03.937831 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:03 crc kubenswrapper[4871]: I1008 00:30:03.948911 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e601ded9-3c4a-4ddf-b5a1-c27eb017e8be" (UID: "e601ded9-3c4a-4ddf-b5a1-c27eb017e8be"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:30:04 crc kubenswrapper[4871]: I1008 00:30:04.040260 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:04 crc kubenswrapper[4871]: I1008 00:30:04.040597 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjzdn\" (UniqueName: \"kubernetes.io/projected/e601ded9-3c4a-4ddf-b5a1-c27eb017e8be-kube-api-access-bjzdn\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:04 crc kubenswrapper[4871]: I1008 00:30:04.356649 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" event={"ID":"e601ded9-3c4a-4ddf-b5a1-c27eb017e8be","Type":"ContainerDied","Data":"49e4866bd7ca593719aca703247dfd64cbd2fe1e7d8b742d270157760d29aff2"} Oct 08 00:30:04 crc kubenswrapper[4871]: I1008 00:30:04.357006 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49e4866bd7ca593719aca703247dfd64cbd2fe1e7d8b742d270157760d29aff2" Oct 08 00:30:04 crc kubenswrapper[4871]: I1008 00:30:04.356746 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-zt7bh" Oct 08 00:30:04 crc kubenswrapper[4871]: I1008 00:30:04.868519 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl"] Oct 08 00:30:04 crc kubenswrapper[4871]: I1008 00:30:04.876819 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331345-4bbkl"] Oct 08 00:30:04 crc kubenswrapper[4871]: I1008 00:30:04.983601 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:30:04 crc kubenswrapper[4871]: E1008 00:30:04.984213 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:30:04 crc kubenswrapper[4871]: I1008 00:30:04.997242 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e55c10e-3977-4d16-9ecf-769afad5ea95" path="/var/lib/kubelet/pods/9e55c10e-3977-4d16-9ecf-769afad5ea95/volumes" Oct 08 00:30:11 crc kubenswrapper[4871]: I1008 00:30:11.772496 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2j6wr"] Oct 08 00:30:11 crc kubenswrapper[4871]: E1008 00:30:11.774212 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e601ded9-3c4a-4ddf-b5a1-c27eb017e8be" containerName="collect-profiles" Oct 08 00:30:11 crc kubenswrapper[4871]: I1008 00:30:11.774261 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="e601ded9-3c4a-4ddf-b5a1-c27eb017e8be" containerName="collect-profiles" Oct 08 00:30:11 crc kubenswrapper[4871]: I1008 00:30:11.774739 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="e601ded9-3c4a-4ddf-b5a1-c27eb017e8be" containerName="collect-profiles" Oct 08 00:30:11 crc kubenswrapper[4871]: I1008 00:30:11.778063 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:11 crc kubenswrapper[4871]: I1008 00:30:11.799339 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2j6wr"] Oct 08 00:30:11 crc kubenswrapper[4871]: I1008 00:30:11.930759 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-utilities\") pod \"redhat-operators-2j6wr\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:11 crc kubenswrapper[4871]: I1008 00:30:11.931112 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-catalog-content\") pod \"redhat-operators-2j6wr\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:11 crc kubenswrapper[4871]: I1008 00:30:11.931237 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5wfz\" (UniqueName: \"kubernetes.io/projected/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-kube-api-access-q5wfz\") pod \"redhat-operators-2j6wr\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:12 crc kubenswrapper[4871]: I1008 00:30:12.032729 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-catalog-content\") pod \"redhat-operators-2j6wr\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:12 crc kubenswrapper[4871]: I1008 00:30:12.033259 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-catalog-content\") pod \"redhat-operators-2j6wr\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:12 crc kubenswrapper[4871]: I1008 00:30:12.033369 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5wfz\" (UniqueName: \"kubernetes.io/projected/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-kube-api-access-q5wfz\") pod \"redhat-operators-2j6wr\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:12 crc kubenswrapper[4871]: I1008 00:30:12.033732 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-utilities\") pod \"redhat-operators-2j6wr\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:12 crc kubenswrapper[4871]: I1008 00:30:12.033963 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-utilities\") pod \"redhat-operators-2j6wr\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:12 crc kubenswrapper[4871]: I1008 00:30:12.054443 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5wfz\" (UniqueName: \"kubernetes.io/projected/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-kube-api-access-q5wfz\") pod \"redhat-operators-2j6wr\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:12 crc kubenswrapper[4871]: I1008 00:30:12.138278 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:12 crc kubenswrapper[4871]: I1008 00:30:12.491875 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2j6wr"] Oct 08 00:30:13 crc kubenswrapper[4871]: I1008 00:30:13.460663 4871 generic.go:334] "Generic (PLEG): container finished" podID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerID="3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6" exitCode=0 Oct 08 00:30:13 crc kubenswrapper[4871]: I1008 00:30:13.460953 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j6wr" event={"ID":"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7","Type":"ContainerDied","Data":"3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6"} Oct 08 00:30:13 crc kubenswrapper[4871]: I1008 00:30:13.460981 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j6wr" event={"ID":"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7","Type":"ContainerStarted","Data":"1d796729a0ab2fd2c34fda53121be4f99665da221c0e56b0ee67ef3637862a7c"} Oct 08 00:30:15 crc kubenswrapper[4871]: I1008 00:30:15.983297 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:30:15 crc kubenswrapper[4871]: E1008 00:30:15.983941 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:30:16 crc kubenswrapper[4871]: I1008 00:30:16.497409 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j6wr" event={"ID":"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7","Type":"ContainerStarted","Data":"93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c"} Oct 08 00:30:18 crc kubenswrapper[4871]: I1008 00:30:18.534127 4871 generic.go:334] "Generic (PLEG): container finished" podID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerID="93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c" exitCode=0 Oct 08 00:30:18 crc kubenswrapper[4871]: I1008 00:30:18.534330 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j6wr" event={"ID":"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7","Type":"ContainerDied","Data":"93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c"} Oct 08 00:30:19 crc kubenswrapper[4871]: I1008 00:30:19.552097 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j6wr" event={"ID":"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7","Type":"ContainerStarted","Data":"5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23"} Oct 08 00:30:19 crc kubenswrapper[4871]: I1008 00:30:19.574006 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2j6wr" podStartSLOduration=3.029970762 podStartE2EDuration="8.573978531s" podCreationTimestamp="2025-10-08 00:30:11 +0000 UTC" firstStartedPulling="2025-10-08 00:30:13.46412414 +0000 UTC m=+8487.266822243" lastFinishedPulling="2025-10-08 00:30:19.008131939 +0000 UTC m=+8492.810830012" observedRunningTime="2025-10-08 00:30:19.573071147 +0000 UTC m=+8493.375769220" watchObservedRunningTime="2025-10-08 00:30:19.573978531 +0000 UTC m=+8493.376676644" Oct 08 00:30:22 crc kubenswrapper[4871]: I1008 00:30:22.138662 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:22 crc kubenswrapper[4871]: I1008 00:30:22.139586 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:23 crc kubenswrapper[4871]: I1008 00:30:23.209344 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2j6wr" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerName="registry-server" probeResult="failure" output=< Oct 08 00:30:23 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 08 00:30:23 crc kubenswrapper[4871]: > Oct 08 00:30:30 crc kubenswrapper[4871]: I1008 00:30:30.982881 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:30:30 crc kubenswrapper[4871]: E1008 00:30:30.983699 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:30:32 crc kubenswrapper[4871]: I1008 00:30:32.214760 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:32 crc kubenswrapper[4871]: I1008 00:30:32.284917 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:32 crc kubenswrapper[4871]: I1008 00:30:32.448440 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2j6wr"] Oct 08 00:30:33 crc kubenswrapper[4871]: I1008 00:30:33.699445 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2j6wr" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerName="registry-server" containerID="cri-o://5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23" gracePeriod=2 Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.280428 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.433638 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5wfz\" (UniqueName: \"kubernetes.io/projected/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-kube-api-access-q5wfz\") pod \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.433677 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-utilities\") pod \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.433801 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-catalog-content\") pod \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\" (UID: \"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7\") " Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.435438 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-utilities" (OuterVolumeSpecName: "utilities") pod "2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" (UID: "2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.440004 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-kube-api-access-q5wfz" (OuterVolumeSpecName: "kube-api-access-q5wfz") pod "2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" (UID: "2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7"). InnerVolumeSpecName "kube-api-access-q5wfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.518677 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" (UID: "2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.536519 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.536559 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5wfz\" (UniqueName: \"kubernetes.io/projected/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-kube-api-access-q5wfz\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.536569 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.714536 4871 generic.go:334] "Generic (PLEG): container finished" podID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerID="5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23" exitCode=0 Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.714588 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j6wr" event={"ID":"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7","Type":"ContainerDied","Data":"5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23"} Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.714624 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j6wr" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.714637 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j6wr" event={"ID":"2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7","Type":"ContainerDied","Data":"1d796729a0ab2fd2c34fda53121be4f99665da221c0e56b0ee67ef3637862a7c"} Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.714666 4871 scope.go:117] "RemoveContainer" containerID="5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.742499 4871 scope.go:117] "RemoveContainer" containerID="93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.765840 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2j6wr"] Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.774628 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2j6wr"] Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.786211 4871 scope.go:117] "RemoveContainer" containerID="3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.829988 4871 scope.go:117] "RemoveContainer" containerID="5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23" Oct 08 00:30:34 crc kubenswrapper[4871]: E1008 00:30:34.830572 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23\": container with ID starting with 5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23 not found: ID does not exist" containerID="5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.830651 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23"} err="failed to get container status \"5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23\": rpc error: code = NotFound desc = could not find container \"5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23\": container with ID starting with 5ce1f663ecff0a5c945bac3969dbed22f15c9ebc914f69e6ae4651f5b612ec23 not found: ID does not exist" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.830678 4871 scope.go:117] "RemoveContainer" containerID="93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c" Oct 08 00:30:34 crc kubenswrapper[4871]: E1008 00:30:34.831159 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c\": container with ID starting with 93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c not found: ID does not exist" containerID="93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.831237 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c"} err="failed to get container status \"93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c\": rpc error: code = NotFound desc = could not find container \"93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c\": container with ID starting with 93b3e7eaf4d5e7fa4228d1d4aec6432ea1879b2caf67b5dd5081b682d3ebba7c not found: ID does not exist" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.831296 4871 scope.go:117] "RemoveContainer" containerID="3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6" Oct 08 00:30:34 crc kubenswrapper[4871]: E1008 00:30:34.832053 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6\": container with ID starting with 3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6 not found: ID does not exist" containerID="3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.832096 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6"} err="failed to get container status \"3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6\": rpc error: code = NotFound desc = could not find container \"3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6\": container with ID starting with 3d00d101e5cbef117c179be2ca18894d81b51b1cea5238850b38f7ddcc40c6e6 not found: ID does not exist" Oct 08 00:30:34 crc kubenswrapper[4871]: I1008 00:30:34.998187 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" path="/var/lib/kubelet/pods/2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7/volumes" Oct 08 00:30:44 crc kubenswrapper[4871]: I1008 00:30:44.982784 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:30:44 crc kubenswrapper[4871]: E1008 00:30:44.983851 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:30:55 crc kubenswrapper[4871]: I1008 00:30:55.982461 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:30:55 crc kubenswrapper[4871]: E1008 00:30:55.983275 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:31:03 crc kubenswrapper[4871]: I1008 00:31:03.316158 4871 scope.go:117] "RemoveContainer" containerID="de57788daf996cb4b12b510173765a62c8ee548135e19b0f772efa3fb5283b40" Oct 08 00:31:08 crc kubenswrapper[4871]: I1008 00:31:08.982812 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:31:08 crc kubenswrapper[4871]: E1008 00:31:08.983538 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:31:19 crc kubenswrapper[4871]: I1008 00:31:19.982917 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:31:19 crc kubenswrapper[4871]: E1008 00:31:19.983739 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:31:32 crc kubenswrapper[4871]: I1008 00:31:32.982997 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:31:32 crc kubenswrapper[4871]: E1008 00:31:32.983923 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:31:46 crc kubenswrapper[4871]: I1008 00:31:46.996931 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:31:47 crc kubenswrapper[4871]: E1008 00:31:47.014225 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:31:58 crc kubenswrapper[4871]: I1008 00:31:58.983109 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:31:58 crc kubenswrapper[4871]: E1008 00:31:58.984032 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:32:09 crc kubenswrapper[4871]: I1008 00:32:09.983543 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:32:09 crc kubenswrapper[4871]: E1008 00:32:09.984371 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:32:22 crc kubenswrapper[4871]: I1008 00:32:22.982947 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:32:22 crc kubenswrapper[4871]: E1008 00:32:22.985071 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:32:33 crc kubenswrapper[4871]: I1008 00:32:33.982693 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:32:33 crc kubenswrapper[4871]: E1008 00:32:33.983708 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:32:47 crc kubenswrapper[4871]: I1008 00:32:47.982561 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:32:47 crc kubenswrapper[4871]: E1008 00:32:47.983337 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:33:00 crc kubenswrapper[4871]: I1008 00:33:00.982613 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:33:00 crc kubenswrapper[4871]: E1008 00:33:00.983597 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:33:12 crc kubenswrapper[4871]: I1008 00:33:12.983352 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:33:12 crc kubenswrapper[4871]: E1008 00:33:12.984327 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:33:23 crc kubenswrapper[4871]: I1008 00:33:23.983015 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:33:23 crc kubenswrapper[4871]: E1008 00:33:23.985112 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:33:34 crc kubenswrapper[4871]: I1008 00:33:34.983151 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:33:34 crc kubenswrapper[4871]: E1008 00:33:34.984021 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:33:48 crc kubenswrapper[4871]: I1008 00:33:48.985576 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:33:48 crc kubenswrapper[4871]: E1008 00:33:48.986304 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:33:59 crc kubenswrapper[4871]: I1008 00:33:59.982964 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:33:59 crc kubenswrapper[4871]: E1008 00:33:59.983867 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:34:03 crc kubenswrapper[4871]: I1008 00:34:03.484566 4871 scope.go:117] "RemoveContainer" containerID="95979ff67ff53abf10499f100d18cf45c6e20cf7ee01fddaf514ad6698accd36" Oct 08 00:34:03 crc kubenswrapper[4871]: I1008 00:34:03.508062 4871 scope.go:117] "RemoveContainer" containerID="41a0c0261fee64744ddb7c9770d6aa8b8b83c9a2b7ccb8ee9c0836349df13a97" Oct 08 00:34:03 crc kubenswrapper[4871]: I1008 00:34:03.545965 4871 scope.go:117] "RemoveContainer" containerID="60b9876d18da8bc0c27df4c9ae0bf086882e6f7a48b7974621fd16be4af23fd0" Oct 08 00:34:13 crc kubenswrapper[4871]: I1008 00:34:13.982937 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:34:13 crc kubenswrapper[4871]: E1008 00:34:13.984191 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:34:27 crc kubenswrapper[4871]: I1008 00:34:27.982870 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:34:27 crc kubenswrapper[4871]: E1008 00:34:27.983757 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:34:33 crc kubenswrapper[4871]: I1008 00:34:33.760006 4871 generic.go:334] "Generic (PLEG): container finished" podID="7b13172c-c2ed-41f2-ab24-538418c2e1b3" containerID="24803e34d3a8239446f1a413315e33e5bd967f2674f4d78e9281bc969e0d9e21" exitCode=0 Oct 08 00:34:33 crc kubenswrapper[4871]: I1008 00:34:33.760038 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" event={"ID":"7b13172c-c2ed-41f2-ab24-538418c2e1b3","Type":"ContainerDied","Data":"24803e34d3a8239446f1a413315e33e5bd967f2674f4d78e9281bc969e0d9e21"} Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.375947 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.447427 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-telemetry-combined-ca-bundle\") pod \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.447562 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-inventory\") pod \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.447642 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-0\") pod \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.447673 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-1\") pod \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.447716 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ssh-key\") pod \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.447828 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p2l8\" (UniqueName: \"kubernetes.io/projected/7b13172c-c2ed-41f2-ab24-538418c2e1b3-kube-api-access-9p2l8\") pod \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.447875 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-2\") pod \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.447927 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceph\") pod \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\" (UID: \"7b13172c-c2ed-41f2-ab24-538418c2e1b3\") " Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.453685 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b13172c-c2ed-41f2-ab24-538418c2e1b3-kube-api-access-9p2l8" (OuterVolumeSpecName: "kube-api-access-9p2l8") pod "7b13172c-c2ed-41f2-ab24-538418c2e1b3" (UID: "7b13172c-c2ed-41f2-ab24-538418c2e1b3"). InnerVolumeSpecName "kube-api-access-9p2l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.453756 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7b13172c-c2ed-41f2-ab24-538418c2e1b3" (UID: "7b13172c-c2ed-41f2-ab24-538418c2e1b3"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.458765 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceph" (OuterVolumeSpecName: "ceph") pod "7b13172c-c2ed-41f2-ab24-538418c2e1b3" (UID: "7b13172c-c2ed-41f2-ab24-538418c2e1b3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.478640 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7b13172c-c2ed-41f2-ab24-538418c2e1b3" (UID: "7b13172c-c2ed-41f2-ab24-538418c2e1b3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.482470 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "7b13172c-c2ed-41f2-ab24-538418c2e1b3" (UID: "7b13172c-c2ed-41f2-ab24-538418c2e1b3"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.491233 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-inventory" (OuterVolumeSpecName: "inventory") pod "7b13172c-c2ed-41f2-ab24-538418c2e1b3" (UID: "7b13172c-c2ed-41f2-ab24-538418c2e1b3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.503199 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "7b13172c-c2ed-41f2-ab24-538418c2e1b3" (UID: "7b13172c-c2ed-41f2-ab24-538418c2e1b3"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.506946 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "7b13172c-c2ed-41f2-ab24-538418c2e1b3" (UID: "7b13172c-c2ed-41f2-ab24-538418c2e1b3"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.550819 4871 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.550868 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.550882 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p2l8\" (UniqueName: \"kubernetes.io/projected/7b13172c-c2ed-41f2-ab24-538418c2e1b3-kube-api-access-9p2l8\") on node \"crc\" DevicePath \"\"" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.550893 4871 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.550909 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.550918 4871 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.550928 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.550938 4871 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7b13172c-c2ed-41f2-ab24-538418c2e1b3-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.786500 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" event={"ID":"7b13172c-c2ed-41f2-ab24-538418c2e1b3","Type":"ContainerDied","Data":"4954327643ffea2cfe7633b99dd5682b27f2d5c4115c22069ad0b0dfcc0b3e21"} Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.786563 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-jmt5b" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.786565 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4954327643ffea2cfe7633b99dd5682b27f2d5c4115c22069ad0b0dfcc0b3e21" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.944523 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-s9zww"] Oct 08 00:34:35 crc kubenswrapper[4871]: E1008 00:34:35.945109 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerName="extract-utilities" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.945129 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerName="extract-utilities" Oct 08 00:34:35 crc kubenswrapper[4871]: E1008 00:34:35.945193 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerName="registry-server" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.945203 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerName="registry-server" Oct 08 00:34:35 crc kubenswrapper[4871]: E1008 00:34:35.945220 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerName="extract-content" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.945239 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerName="extract-content" Oct 08 00:34:35 crc kubenswrapper[4871]: E1008 00:34:35.945256 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b13172c-c2ed-41f2-ab24-538418c2e1b3" containerName="telemetry-openstack-openstack-cell1" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.945266 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b13172c-c2ed-41f2-ab24-538418c2e1b3" containerName="telemetry-openstack-openstack-cell1" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.945519 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b13172c-c2ed-41f2-ab24-538418c2e1b3" containerName="telemetry-openstack-openstack-cell1" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.945554 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eb50d8c-dd3e-42ad-9b74-33a9dddbe9c7" containerName="registry-server" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.946476 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.949486 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.949601 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.949818 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.949909 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.951525 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.955680 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-s9zww"] Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.961559 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.961644 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgwzw\" (UniqueName: \"kubernetes.io/projected/59d68a40-edd5-4599-a9dc-5fa134a4a05c-kube-api-access-tgwzw\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.961699 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.961734 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.961751 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:35 crc kubenswrapper[4871]: I1008 00:34:35.961957 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.064882 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.065169 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.065246 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgwzw\" (UniqueName: \"kubernetes.io/projected/59d68a40-edd5-4599-a9dc-5fa134a4a05c-kube-api-access-tgwzw\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.065313 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.065383 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.065422 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.068784 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.070746 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.071693 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.076703 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.078082 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.082543 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgwzw\" (UniqueName: \"kubernetes.io/projected/59d68a40-edd5-4599-a9dc-5fa134a4a05c-kube-api-access-tgwzw\") pod \"neutron-sriov-openstack-openstack-cell1-s9zww\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.272909 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:34:36 crc kubenswrapper[4871]: I1008 00:34:36.862535 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-s9zww"] Oct 08 00:34:37 crc kubenswrapper[4871]: I1008 00:34:37.816644 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" event={"ID":"59d68a40-edd5-4599-a9dc-5fa134a4a05c","Type":"ContainerStarted","Data":"f9471ed0653c8f38e4a9765444b7a31f72238f6a5735ea0a49f6144c410cccb8"} Oct 08 00:34:37 crc kubenswrapper[4871]: I1008 00:34:37.817322 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" event={"ID":"59d68a40-edd5-4599-a9dc-5fa134a4a05c","Type":"ContainerStarted","Data":"7628c513b1c265af4315a32ce139e877ea3d3f326c54c4f6a7e7b94646aedfc6"} Oct 08 00:34:37 crc kubenswrapper[4871]: I1008 00:34:37.850852 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" podStartSLOduration=2.323143523 podStartE2EDuration="2.850818869s" podCreationTimestamp="2025-10-08 00:34:35 +0000 UTC" firstStartedPulling="2025-10-08 00:34:36.867288633 +0000 UTC m=+8750.669986706" lastFinishedPulling="2025-10-08 00:34:37.394963939 +0000 UTC m=+8751.197662052" observedRunningTime="2025-10-08 00:34:37.845141391 +0000 UTC m=+8751.647839514" watchObservedRunningTime="2025-10-08 00:34:37.850818869 +0000 UTC m=+8751.653516982" Oct 08 00:34:40 crc kubenswrapper[4871]: I1008 00:34:40.985751 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:34:41 crc kubenswrapper[4871]: I1008 00:34:41.887567 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"8dc7becc245525b7cc3f6ec5e629e5e2e2a65d184a4803a875a211defc59af39"} Oct 08 00:37:05 crc kubenswrapper[4871]: I1008 00:37:05.512287 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:37:05 crc kubenswrapper[4871]: I1008 00:37:05.513038 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.174024 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ljsz7"] Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.178077 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.196383 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ljsz7"] Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.347056 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-utilities\") pod \"certified-operators-ljsz7\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.348011 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmhxz\" (UniqueName: \"kubernetes.io/projected/c797e852-e457-4b40-9536-9c689a1d6320-kube-api-access-wmhxz\") pod \"certified-operators-ljsz7\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.348142 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-catalog-content\") pod \"certified-operators-ljsz7\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.450055 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-utilities\") pod \"certified-operators-ljsz7\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.450252 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmhxz\" (UniqueName: \"kubernetes.io/projected/c797e852-e457-4b40-9536-9c689a1d6320-kube-api-access-wmhxz\") pod \"certified-operators-ljsz7\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.450307 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-catalog-content\") pod \"certified-operators-ljsz7\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.450820 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-catalog-content\") pod \"certified-operators-ljsz7\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.450992 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-utilities\") pod \"certified-operators-ljsz7\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.473390 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmhxz\" (UniqueName: \"kubernetes.io/projected/c797e852-e457-4b40-9536-9c689a1d6320-kube-api-access-wmhxz\") pod \"certified-operators-ljsz7\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:14 crc kubenswrapper[4871]: I1008 00:37:14.540781 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:15 crc kubenswrapper[4871]: I1008 00:37:15.058174 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ljsz7"] Oct 08 00:37:15 crc kubenswrapper[4871]: W1008 00:37:15.063823 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc797e852_e457_4b40_9536_9c689a1d6320.slice/crio-445ec09c60cda3540fac8172bba701d72c0f4abc24f46a5a6bb1603e7ee7c80e WatchSource:0}: Error finding container 445ec09c60cda3540fac8172bba701d72c0f4abc24f46a5a6bb1603e7ee7c80e: Status 404 returned error can't find the container with id 445ec09c60cda3540fac8172bba701d72c0f4abc24f46a5a6bb1603e7ee7c80e Oct 08 00:37:15 crc kubenswrapper[4871]: I1008 00:37:15.733772 4871 generic.go:334] "Generic (PLEG): container finished" podID="c797e852-e457-4b40-9536-9c689a1d6320" containerID="c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77" exitCode=0 Oct 08 00:37:15 crc kubenswrapper[4871]: I1008 00:37:15.733937 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljsz7" event={"ID":"c797e852-e457-4b40-9536-9c689a1d6320","Type":"ContainerDied","Data":"c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77"} Oct 08 00:37:15 crc kubenswrapper[4871]: I1008 00:37:15.734139 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljsz7" event={"ID":"c797e852-e457-4b40-9536-9c689a1d6320","Type":"ContainerStarted","Data":"445ec09c60cda3540fac8172bba701d72c0f4abc24f46a5a6bb1603e7ee7c80e"} Oct 08 00:37:15 crc kubenswrapper[4871]: I1008 00:37:15.737934 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:37:16 crc kubenswrapper[4871]: I1008 00:37:16.756378 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljsz7" event={"ID":"c797e852-e457-4b40-9536-9c689a1d6320","Type":"ContainerStarted","Data":"b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72"} Oct 08 00:37:17 crc kubenswrapper[4871]: I1008 00:37:17.770492 4871 generic.go:334] "Generic (PLEG): container finished" podID="c797e852-e457-4b40-9536-9c689a1d6320" containerID="b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72" exitCode=0 Oct 08 00:37:17 crc kubenswrapper[4871]: I1008 00:37:17.770557 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljsz7" event={"ID":"c797e852-e457-4b40-9536-9c689a1d6320","Type":"ContainerDied","Data":"b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72"} Oct 08 00:37:18 crc kubenswrapper[4871]: I1008 00:37:18.783342 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljsz7" event={"ID":"c797e852-e457-4b40-9536-9c689a1d6320","Type":"ContainerStarted","Data":"eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef"} Oct 08 00:37:18 crc kubenswrapper[4871]: I1008 00:37:18.803415 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ljsz7" podStartSLOduration=2.36058596 podStartE2EDuration="4.803389961s" podCreationTimestamp="2025-10-08 00:37:14 +0000 UTC" firstStartedPulling="2025-10-08 00:37:15.737530765 +0000 UTC m=+8909.540228848" lastFinishedPulling="2025-10-08 00:37:18.180334736 +0000 UTC m=+8911.983032849" observedRunningTime="2025-10-08 00:37:18.799170561 +0000 UTC m=+8912.601868644" watchObservedRunningTime="2025-10-08 00:37:18.803389961 +0000 UTC m=+8912.606088054" Oct 08 00:37:24 crc kubenswrapper[4871]: I1008 00:37:24.541177 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:24 crc kubenswrapper[4871]: I1008 00:37:24.541675 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:24 crc kubenswrapper[4871]: I1008 00:37:24.595270 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:24 crc kubenswrapper[4871]: I1008 00:37:24.902597 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:24 crc kubenswrapper[4871]: I1008 00:37:24.964174 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ljsz7"] Oct 08 00:37:26 crc kubenswrapper[4871]: I1008 00:37:26.870061 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ljsz7" podUID="c797e852-e457-4b40-9536-9c689a1d6320" containerName="registry-server" containerID="cri-o://eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef" gracePeriod=2 Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.390881 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.555421 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-catalog-content\") pod \"c797e852-e457-4b40-9536-9c689a1d6320\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.555679 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-utilities\") pod \"c797e852-e457-4b40-9536-9c689a1d6320\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.555825 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmhxz\" (UniqueName: \"kubernetes.io/projected/c797e852-e457-4b40-9536-9c689a1d6320-kube-api-access-wmhxz\") pod \"c797e852-e457-4b40-9536-9c689a1d6320\" (UID: \"c797e852-e457-4b40-9536-9c689a1d6320\") " Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.556470 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-utilities" (OuterVolumeSpecName: "utilities") pod "c797e852-e457-4b40-9536-9c689a1d6320" (UID: "c797e852-e457-4b40-9536-9c689a1d6320"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.556872 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.568944 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c797e852-e457-4b40-9536-9c689a1d6320-kube-api-access-wmhxz" (OuterVolumeSpecName: "kube-api-access-wmhxz") pod "c797e852-e457-4b40-9536-9c689a1d6320" (UID: "c797e852-e457-4b40-9536-9c689a1d6320"). InnerVolumeSpecName "kube-api-access-wmhxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.603554 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c797e852-e457-4b40-9536-9c689a1d6320" (UID: "c797e852-e457-4b40-9536-9c689a1d6320"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.658996 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmhxz\" (UniqueName: \"kubernetes.io/projected/c797e852-e457-4b40-9536-9c689a1d6320-kube-api-access-wmhxz\") on node \"crc\" DevicePath \"\"" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.659028 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c797e852-e457-4b40-9536-9c689a1d6320-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.884731 4871 generic.go:334] "Generic (PLEG): container finished" podID="c797e852-e457-4b40-9536-9c689a1d6320" containerID="eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef" exitCode=0 Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.884862 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ljsz7" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.884870 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljsz7" event={"ID":"c797e852-e457-4b40-9536-9c689a1d6320","Type":"ContainerDied","Data":"eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef"} Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.885313 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljsz7" event={"ID":"c797e852-e457-4b40-9536-9c689a1d6320","Type":"ContainerDied","Data":"445ec09c60cda3540fac8172bba701d72c0f4abc24f46a5a6bb1603e7ee7c80e"} Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.885347 4871 scope.go:117] "RemoveContainer" containerID="eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.934217 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ljsz7"] Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.935777 4871 scope.go:117] "RemoveContainer" containerID="b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72" Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.946420 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ljsz7"] Oct 08 00:37:27 crc kubenswrapper[4871]: I1008 00:37:27.966989 4871 scope.go:117] "RemoveContainer" containerID="c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77" Oct 08 00:37:28 crc kubenswrapper[4871]: I1008 00:37:28.054521 4871 scope.go:117] "RemoveContainer" containerID="eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef" Oct 08 00:37:28 crc kubenswrapper[4871]: E1008 00:37:28.055467 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef\": container with ID starting with eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef not found: ID does not exist" containerID="eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef" Oct 08 00:37:28 crc kubenswrapper[4871]: I1008 00:37:28.055499 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef"} err="failed to get container status \"eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef\": rpc error: code = NotFound desc = could not find container \"eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef\": container with ID starting with eef6049d25e9520b7268e2bc88871f2365507d01ccb37772c7ebba2e3b738fef not found: ID does not exist" Oct 08 00:37:28 crc kubenswrapper[4871]: I1008 00:37:28.055618 4871 scope.go:117] "RemoveContainer" containerID="b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72" Oct 08 00:37:28 crc kubenswrapper[4871]: E1008 00:37:28.056077 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72\": container with ID starting with b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72 not found: ID does not exist" containerID="b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72" Oct 08 00:37:28 crc kubenswrapper[4871]: I1008 00:37:28.056101 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72"} err="failed to get container status \"b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72\": rpc error: code = NotFound desc = could not find container \"b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72\": container with ID starting with b84cbc401afacc194d7f973353c3dca93de2f778374b5ad08acfd2b2f9776e72 not found: ID does not exist" Oct 08 00:37:28 crc kubenswrapper[4871]: I1008 00:37:28.056119 4871 scope.go:117] "RemoveContainer" containerID="c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77" Oct 08 00:37:28 crc kubenswrapper[4871]: E1008 00:37:28.057064 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77\": container with ID starting with c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77 not found: ID does not exist" containerID="c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77" Oct 08 00:37:28 crc kubenswrapper[4871]: I1008 00:37:28.057093 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77"} err="failed to get container status \"c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77\": rpc error: code = NotFound desc = could not find container \"c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77\": container with ID starting with c5e38386134b1832623eadcbef487c68a08d5d3d964d00d46b7c9bbd55dfcb77 not found: ID does not exist" Oct 08 00:37:29 crc kubenswrapper[4871]: I1008 00:37:29.001409 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c797e852-e457-4b40-9536-9c689a1d6320" path="/var/lib/kubelet/pods/c797e852-e457-4b40-9536-9c689a1d6320/volumes" Oct 08 00:37:35 crc kubenswrapper[4871]: I1008 00:37:35.511944 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:37:35 crc kubenswrapper[4871]: I1008 00:37:35.512612 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:38:05 crc kubenswrapper[4871]: I1008 00:38:05.512618 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:38:05 crc kubenswrapper[4871]: I1008 00:38:05.513538 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:38:05 crc kubenswrapper[4871]: I1008 00:38:05.513618 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:38:05 crc kubenswrapper[4871]: I1008 00:38:05.515050 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8dc7becc245525b7cc3f6ec5e629e5e2e2a65d184a4803a875a211defc59af39"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:38:05 crc kubenswrapper[4871]: I1008 00:38:05.515184 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://8dc7becc245525b7cc3f6ec5e629e5e2e2a65d184a4803a875a211defc59af39" gracePeriod=600 Oct 08 00:38:06 crc kubenswrapper[4871]: I1008 00:38:06.379924 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="8dc7becc245525b7cc3f6ec5e629e5e2e2a65d184a4803a875a211defc59af39" exitCode=0 Oct 08 00:38:06 crc kubenswrapper[4871]: I1008 00:38:06.379984 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"8dc7becc245525b7cc3f6ec5e629e5e2e2a65d184a4803a875a211defc59af39"} Oct 08 00:38:06 crc kubenswrapper[4871]: I1008 00:38:06.380524 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877"} Oct 08 00:38:06 crc kubenswrapper[4871]: I1008 00:38:06.380547 4871 scope.go:117] "RemoveContainer" containerID="b8ebb86490f6bf5556249f8dd24de6eb400d31d9fec08e4617610ab5174b8c9d" Oct 08 00:38:14 crc kubenswrapper[4871]: I1008 00:38:14.489623 4871 generic.go:334] "Generic (PLEG): container finished" podID="59d68a40-edd5-4599-a9dc-5fa134a4a05c" containerID="f9471ed0653c8f38e4a9765444b7a31f72238f6a5735ea0a49f6144c410cccb8" exitCode=0 Oct 08 00:38:14 crc kubenswrapper[4871]: I1008 00:38:14.489733 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" event={"ID":"59d68a40-edd5-4599-a9dc-5fa134a4a05c","Type":"ContainerDied","Data":"f9471ed0653c8f38e4a9765444b7a31f72238f6a5735ea0a49f6144c410cccb8"} Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.081132 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.167132 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-inventory\") pod \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.167298 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgwzw\" (UniqueName: \"kubernetes.io/projected/59d68a40-edd5-4599-a9dc-5fa134a4a05c-kube-api-access-tgwzw\") pod \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.167445 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-combined-ca-bundle\") pod \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.167586 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ceph\") pod \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.167724 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-agent-neutron-config-0\") pod \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.167769 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ssh-key\") pod \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\" (UID: \"59d68a40-edd5-4599-a9dc-5fa134a4a05c\") " Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.176235 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "59d68a40-edd5-4599-a9dc-5fa134a4a05c" (UID: "59d68a40-edd5-4599-a9dc-5fa134a4a05c"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.176260 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ceph" (OuterVolumeSpecName: "ceph") pod "59d68a40-edd5-4599-a9dc-5fa134a4a05c" (UID: "59d68a40-edd5-4599-a9dc-5fa134a4a05c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.177444 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59d68a40-edd5-4599-a9dc-5fa134a4a05c-kube-api-access-tgwzw" (OuterVolumeSpecName: "kube-api-access-tgwzw") pod "59d68a40-edd5-4599-a9dc-5fa134a4a05c" (UID: "59d68a40-edd5-4599-a9dc-5fa134a4a05c"). InnerVolumeSpecName "kube-api-access-tgwzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.202131 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "59d68a40-edd5-4599-a9dc-5fa134a4a05c" (UID: "59d68a40-edd5-4599-a9dc-5fa134a4a05c"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.202909 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-inventory" (OuterVolumeSpecName: "inventory") pod "59d68a40-edd5-4599-a9dc-5fa134a4a05c" (UID: "59d68a40-edd5-4599-a9dc-5fa134a4a05c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.215172 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "59d68a40-edd5-4599-a9dc-5fa134a4a05c" (UID: "59d68a40-edd5-4599-a9dc-5fa134a4a05c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.271240 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgwzw\" (UniqueName: \"kubernetes.io/projected/59d68a40-edd5-4599-a9dc-5fa134a4a05c-kube-api-access-tgwzw\") on node \"crc\" DevicePath \"\"" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.271293 4871 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.271309 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.271322 4871 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.271334 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.271362 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59d68a40-edd5-4599-a9dc-5fa134a4a05c-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.511561 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" event={"ID":"59d68a40-edd5-4599-a9dc-5fa134a4a05c","Type":"ContainerDied","Data":"7628c513b1c265af4315a32ce139e877ea3d3f326c54c4f6a7e7b94646aedfc6"} Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.511889 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7628c513b1c265af4315a32ce139e877ea3d3f326c54c4f6a7e7b94646aedfc6" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.511692 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-s9zww" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.644217 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs"] Oct 08 00:38:16 crc kubenswrapper[4871]: E1008 00:38:16.645094 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c797e852-e457-4b40-9536-9c689a1d6320" containerName="registry-server" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.645115 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c797e852-e457-4b40-9536-9c689a1d6320" containerName="registry-server" Oct 08 00:38:16 crc kubenswrapper[4871]: E1008 00:38:16.645149 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c797e852-e457-4b40-9536-9c689a1d6320" containerName="extract-content" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.645158 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c797e852-e457-4b40-9536-9c689a1d6320" containerName="extract-content" Oct 08 00:38:16 crc kubenswrapper[4871]: E1008 00:38:16.645189 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c797e852-e457-4b40-9536-9c689a1d6320" containerName="extract-utilities" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.645196 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="c797e852-e457-4b40-9536-9c689a1d6320" containerName="extract-utilities" Oct 08 00:38:16 crc kubenswrapper[4871]: E1008 00:38:16.645224 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d68a40-edd5-4599-a9dc-5fa134a4a05c" containerName="neutron-sriov-openstack-openstack-cell1" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.645232 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d68a40-edd5-4599-a9dc-5fa134a4a05c" containerName="neutron-sriov-openstack-openstack-cell1" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.645511 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="c797e852-e457-4b40-9536-9c689a1d6320" containerName="registry-server" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.645524 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d68a40-edd5-4599-a9dc-5fa134a4a05c" containerName="neutron-sriov-openstack-openstack-cell1" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.646911 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.649466 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.649527 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.649750 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.651129 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.651151 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.660748 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs"] Oct 08 00:38:16 crc kubenswrapper[4871]: E1008 00:38:16.780684 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59d68a40_edd5_4599_a9dc_5fa134a4a05c.slice\": RecentStats: unable to find data in memory cache]" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.785436 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.785520 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.785861 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brxps\" (UniqueName: \"kubernetes.io/projected/ac814cee-d737-4c1f-b486-93d99024c197-kube-api-access-brxps\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.786031 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.786079 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.786103 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.888683 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brxps\" (UniqueName: \"kubernetes.io/projected/ac814cee-d737-4c1f-b486-93d99024c197-kube-api-access-brxps\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.888776 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.888835 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.888855 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.888914 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.888955 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.894240 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.895460 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.895663 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.896093 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.903030 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.914759 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brxps\" (UniqueName: \"kubernetes.io/projected/ac814cee-d737-4c1f-b486-93d99024c197-kube-api-access-brxps\") pod \"neutron-dhcp-openstack-openstack-cell1-rfmrs\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:16 crc kubenswrapper[4871]: I1008 00:38:16.966925 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:38:17 crc kubenswrapper[4871]: I1008 00:38:17.569535 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs"] Oct 08 00:38:18 crc kubenswrapper[4871]: I1008 00:38:18.561919 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" event={"ID":"ac814cee-d737-4c1f-b486-93d99024c197","Type":"ContainerStarted","Data":"167c976dd15143d9cb0f182cc23579aaad3e4f8aa514f7a579b71a7904b595fc"} Oct 08 00:38:18 crc kubenswrapper[4871]: I1008 00:38:18.562278 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" event={"ID":"ac814cee-d737-4c1f-b486-93d99024c197","Type":"ContainerStarted","Data":"dc448799a0acb23d80a19d7603db8a6d698d68c010f1911b53a650b2978bd35a"} Oct 08 00:38:18 crc kubenswrapper[4871]: I1008 00:38:18.595980 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" podStartSLOduration=2.067742034 podStartE2EDuration="2.595960254s" podCreationTimestamp="2025-10-08 00:38:16 +0000 UTC" firstStartedPulling="2025-10-08 00:38:17.581335096 +0000 UTC m=+8971.384033209" lastFinishedPulling="2025-10-08 00:38:18.109553346 +0000 UTC m=+8971.912251429" observedRunningTime="2025-10-08 00:38:18.587910493 +0000 UTC m=+8972.390608566" watchObservedRunningTime="2025-10-08 00:38:18.595960254 +0000 UTC m=+8972.398658327" Oct 08 00:40:05 crc kubenswrapper[4871]: I1008 00:40:05.512354 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:40:05 crc kubenswrapper[4871]: I1008 00:40:05.513101 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:40:35 crc kubenswrapper[4871]: I1008 00:40:35.512419 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:40:35 crc kubenswrapper[4871]: I1008 00:40:35.513114 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.116363 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qvwvn"] Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.120430 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.134073 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qvwvn"] Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.263470 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-utilities\") pod \"redhat-operators-qvwvn\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.265187 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qptjj\" (UniqueName: \"kubernetes.io/projected/f278bee2-1527-494f-bc46-703e08bc91c1-kube-api-access-qptjj\") pod \"redhat-operators-qvwvn\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.265439 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-catalog-content\") pod \"redhat-operators-qvwvn\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.368416 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-utilities\") pod \"redhat-operators-qvwvn\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.368766 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qptjj\" (UniqueName: \"kubernetes.io/projected/f278bee2-1527-494f-bc46-703e08bc91c1-kube-api-access-qptjj\") pod \"redhat-operators-qvwvn\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.368932 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-catalog-content\") pod \"redhat-operators-qvwvn\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.369359 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-utilities\") pod \"redhat-operators-qvwvn\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.369378 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-catalog-content\") pod \"redhat-operators-qvwvn\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.395021 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qptjj\" (UniqueName: \"kubernetes.io/projected/f278bee2-1527-494f-bc46-703e08bc91c1-kube-api-access-qptjj\") pod \"redhat-operators-qvwvn\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.453870 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:04 crc kubenswrapper[4871]: I1008 00:41:04.956721 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qvwvn"] Oct 08 00:41:05 crc kubenswrapper[4871]: I1008 00:41:05.512367 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:41:05 crc kubenswrapper[4871]: I1008 00:41:05.512724 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:41:05 crc kubenswrapper[4871]: I1008 00:41:05.512780 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:41:05 crc kubenswrapper[4871]: I1008 00:41:05.513995 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:41:05 crc kubenswrapper[4871]: I1008 00:41:05.514101 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" gracePeriod=600 Oct 08 00:41:05 crc kubenswrapper[4871]: I1008 00:41:05.576928 4871 generic.go:334] "Generic (PLEG): container finished" podID="f278bee2-1527-494f-bc46-703e08bc91c1" containerID="c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643" exitCode=0 Oct 08 00:41:05 crc kubenswrapper[4871]: I1008 00:41:05.576990 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvwvn" event={"ID":"f278bee2-1527-494f-bc46-703e08bc91c1","Type":"ContainerDied","Data":"c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643"} Oct 08 00:41:05 crc kubenswrapper[4871]: I1008 00:41:05.577029 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvwvn" event={"ID":"f278bee2-1527-494f-bc46-703e08bc91c1","Type":"ContainerStarted","Data":"ae524e4de690aa48c293f567ccbed2d9759594ed7cd8075a8f40d6ed88ea8892"} Oct 08 00:41:05 crc kubenswrapper[4871]: E1008 00:41:05.642709 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:41:06 crc kubenswrapper[4871]: I1008 00:41:06.593755 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" exitCode=0 Oct 08 00:41:06 crc kubenswrapper[4871]: I1008 00:41:06.593891 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877"} Oct 08 00:41:06 crc kubenswrapper[4871]: I1008 00:41:06.595458 4871 scope.go:117] "RemoveContainer" containerID="8dc7becc245525b7cc3f6ec5e629e5e2e2a65d184a4803a875a211defc59af39" Oct 08 00:41:06 crc kubenswrapper[4871]: I1008 00:41:06.596143 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:41:06 crc kubenswrapper[4871]: E1008 00:41:06.596676 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:41:07 crc kubenswrapper[4871]: I1008 00:41:07.607783 4871 generic.go:334] "Generic (PLEG): container finished" podID="f278bee2-1527-494f-bc46-703e08bc91c1" containerID="822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588" exitCode=0 Oct 08 00:41:07 crc kubenswrapper[4871]: I1008 00:41:07.607926 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvwvn" event={"ID":"f278bee2-1527-494f-bc46-703e08bc91c1","Type":"ContainerDied","Data":"822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588"} Oct 08 00:41:08 crc kubenswrapper[4871]: I1008 00:41:08.621938 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvwvn" event={"ID":"f278bee2-1527-494f-bc46-703e08bc91c1","Type":"ContainerStarted","Data":"d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82"} Oct 08 00:41:08 crc kubenswrapper[4871]: I1008 00:41:08.653462 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qvwvn" podStartSLOduration=2.219845972 podStartE2EDuration="4.653435983s" podCreationTimestamp="2025-10-08 00:41:04 +0000 UTC" firstStartedPulling="2025-10-08 00:41:05.579566598 +0000 UTC m=+9139.382264671" lastFinishedPulling="2025-10-08 00:41:08.013156579 +0000 UTC m=+9141.815854682" observedRunningTime="2025-10-08 00:41:08.639408737 +0000 UTC m=+9142.442106840" watchObservedRunningTime="2025-10-08 00:41:08.653435983 +0000 UTC m=+9142.456134096" Oct 08 00:41:14 crc kubenswrapper[4871]: I1008 00:41:14.455309 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:14 crc kubenswrapper[4871]: I1008 00:41:14.455828 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:14 crc kubenswrapper[4871]: I1008 00:41:14.507508 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:14 crc kubenswrapper[4871]: I1008 00:41:14.737156 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:14 crc kubenswrapper[4871]: I1008 00:41:14.791507 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qvwvn"] Oct 08 00:41:16 crc kubenswrapper[4871]: I1008 00:41:16.712176 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qvwvn" podUID="f278bee2-1527-494f-bc46-703e08bc91c1" containerName="registry-server" containerID="cri-o://d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82" gracePeriod=2 Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.271458 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.281314 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qptjj\" (UniqueName: \"kubernetes.io/projected/f278bee2-1527-494f-bc46-703e08bc91c1-kube-api-access-qptjj\") pod \"f278bee2-1527-494f-bc46-703e08bc91c1\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.281382 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-utilities\") pod \"f278bee2-1527-494f-bc46-703e08bc91c1\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.281445 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-catalog-content\") pod \"f278bee2-1527-494f-bc46-703e08bc91c1\" (UID: \"f278bee2-1527-494f-bc46-703e08bc91c1\") " Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.284189 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-utilities" (OuterVolumeSpecName: "utilities") pod "f278bee2-1527-494f-bc46-703e08bc91c1" (UID: "f278bee2-1527-494f-bc46-703e08bc91c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.289981 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f278bee2-1527-494f-bc46-703e08bc91c1-kube-api-access-qptjj" (OuterVolumeSpecName: "kube-api-access-qptjj") pod "f278bee2-1527-494f-bc46-703e08bc91c1" (UID: "f278bee2-1527-494f-bc46-703e08bc91c1"). InnerVolumeSpecName "kube-api-access-qptjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.372125 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f278bee2-1527-494f-bc46-703e08bc91c1" (UID: "f278bee2-1527-494f-bc46-703e08bc91c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.383595 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.383639 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qptjj\" (UniqueName: \"kubernetes.io/projected/f278bee2-1527-494f-bc46-703e08bc91c1-kube-api-access-qptjj\") on node \"crc\" DevicePath \"\"" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.383661 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278bee2-1527-494f-bc46-703e08bc91c1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.726049 4871 generic.go:334] "Generic (PLEG): container finished" podID="f278bee2-1527-494f-bc46-703e08bc91c1" containerID="d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82" exitCode=0 Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.726113 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvwvn" event={"ID":"f278bee2-1527-494f-bc46-703e08bc91c1","Type":"ContainerDied","Data":"d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82"} Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.726139 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qvwvn" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.726406 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvwvn" event={"ID":"f278bee2-1527-494f-bc46-703e08bc91c1","Type":"ContainerDied","Data":"ae524e4de690aa48c293f567ccbed2d9759594ed7cd8075a8f40d6ed88ea8892"} Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.726449 4871 scope.go:117] "RemoveContainer" containerID="d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.762717 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qvwvn"] Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.767614 4871 scope.go:117] "RemoveContainer" containerID="822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.783703 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qvwvn"] Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.799474 4871 scope.go:117] "RemoveContainer" containerID="c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.844826 4871 scope.go:117] "RemoveContainer" containerID="d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82" Oct 08 00:41:17 crc kubenswrapper[4871]: E1008 00:41:17.845296 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82\": container with ID starting with d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82 not found: ID does not exist" containerID="d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.845369 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82"} err="failed to get container status \"d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82\": rpc error: code = NotFound desc = could not find container \"d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82\": container with ID starting with d85b75de8c516fe970cf972e4eeb4a7e37877895af861cc444543cbb150c2a82 not found: ID does not exist" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.845395 4871 scope.go:117] "RemoveContainer" containerID="822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588" Oct 08 00:41:17 crc kubenswrapper[4871]: E1008 00:41:17.845707 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588\": container with ID starting with 822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588 not found: ID does not exist" containerID="822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.845744 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588"} err="failed to get container status \"822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588\": rpc error: code = NotFound desc = could not find container \"822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588\": container with ID starting with 822a3c6e9650f13bcc94b5225f85d6533ccb478c874a94d5bdf494ada6cdb588 not found: ID does not exist" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.845771 4871 scope.go:117] "RemoveContainer" containerID="c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643" Oct 08 00:41:17 crc kubenswrapper[4871]: E1008 00:41:17.845983 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643\": container with ID starting with c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643 not found: ID does not exist" containerID="c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643" Oct 08 00:41:17 crc kubenswrapper[4871]: I1008 00:41:17.846012 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643"} err="failed to get container status \"c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643\": rpc error: code = NotFound desc = could not find container \"c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643\": container with ID starting with c07792a35dbdb184f5cf30147a3d87667c47d6c63d8ebc01d0d2c1c37efd6643 not found: ID does not exist" Oct 08 00:41:19 crc kubenswrapper[4871]: I1008 00:41:19.000340 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f278bee2-1527-494f-bc46-703e08bc91c1" path="/var/lib/kubelet/pods/f278bee2-1527-494f-bc46-703e08bc91c1/volumes" Oct 08 00:41:19 crc kubenswrapper[4871]: I1008 00:41:19.983144 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:41:19 crc kubenswrapper[4871]: E1008 00:41:19.984018 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:41:32 crc kubenswrapper[4871]: I1008 00:41:32.983905 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:41:32 crc kubenswrapper[4871]: E1008 00:41:32.984883 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:41:44 crc kubenswrapper[4871]: I1008 00:41:44.983037 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:41:44 crc kubenswrapper[4871]: E1008 00:41:44.983921 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:41:57 crc kubenswrapper[4871]: I1008 00:41:57.982348 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:41:57 crc kubenswrapper[4871]: E1008 00:41:57.983839 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:42:11 crc kubenswrapper[4871]: I1008 00:42:11.984147 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:42:11 crc kubenswrapper[4871]: E1008 00:42:11.985129 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:42:24 crc kubenswrapper[4871]: I1008 00:42:24.983040 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:42:24 crc kubenswrapper[4871]: E1008 00:42:24.984280 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:42:36 crc kubenswrapper[4871]: I1008 00:42:36.989783 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:42:36 crc kubenswrapper[4871]: E1008 00:42:36.990957 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.606716 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fdf5s"] Oct 08 00:42:44 crc kubenswrapper[4871]: E1008 00:42:44.607664 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f278bee2-1527-494f-bc46-703e08bc91c1" containerName="extract-utilities" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.607676 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f278bee2-1527-494f-bc46-703e08bc91c1" containerName="extract-utilities" Oct 08 00:42:44 crc kubenswrapper[4871]: E1008 00:42:44.607692 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f278bee2-1527-494f-bc46-703e08bc91c1" containerName="registry-server" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.607698 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f278bee2-1527-494f-bc46-703e08bc91c1" containerName="registry-server" Oct 08 00:42:44 crc kubenswrapper[4871]: E1008 00:42:44.607730 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f278bee2-1527-494f-bc46-703e08bc91c1" containerName="extract-content" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.607736 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f278bee2-1527-494f-bc46-703e08bc91c1" containerName="extract-content" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.607979 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f278bee2-1527-494f-bc46-703e08bc91c1" containerName="registry-server" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.609478 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.621569 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fdf5s"] Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.767806 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-catalog-content\") pod \"community-operators-fdf5s\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.768181 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kklc\" (UniqueName: \"kubernetes.io/projected/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-kube-api-access-4kklc\") pod \"community-operators-fdf5s\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.768368 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-utilities\") pod \"community-operators-fdf5s\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.871990 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-utilities\") pod \"community-operators-fdf5s\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.872111 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-catalog-content\") pod \"community-operators-fdf5s\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.872145 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kklc\" (UniqueName: \"kubernetes.io/projected/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-kube-api-access-4kklc\") pod \"community-operators-fdf5s\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.874248 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-utilities\") pod \"community-operators-fdf5s\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.874419 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-catalog-content\") pod \"community-operators-fdf5s\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.894596 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kklc\" (UniqueName: \"kubernetes.io/projected/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-kube-api-access-4kklc\") pod \"community-operators-fdf5s\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:44 crc kubenswrapper[4871]: I1008 00:42:44.946190 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:45 crc kubenswrapper[4871]: I1008 00:42:45.552455 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fdf5s"] Oct 08 00:42:45 crc kubenswrapper[4871]: I1008 00:42:45.837699 4871 generic.go:334] "Generic (PLEG): container finished" podID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerID="1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1" exitCode=0 Oct 08 00:42:45 crc kubenswrapper[4871]: I1008 00:42:45.837986 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdf5s" event={"ID":"2e10f5ec-3a61-49f6-a7d4-bf540878b04e","Type":"ContainerDied","Data":"1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1"} Oct 08 00:42:45 crc kubenswrapper[4871]: I1008 00:42:45.838078 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdf5s" event={"ID":"2e10f5ec-3a61-49f6-a7d4-bf540878b04e","Type":"ContainerStarted","Data":"701f24397b8551babe3a6bb84e97de67e0e14cc910171d1b38d885bee5806f36"} Oct 08 00:42:45 crc kubenswrapper[4871]: I1008 00:42:45.840084 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:42:46 crc kubenswrapper[4871]: I1008 00:42:46.851633 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdf5s" event={"ID":"2e10f5ec-3a61-49f6-a7d4-bf540878b04e","Type":"ContainerStarted","Data":"bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96"} Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.008676 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zmvcb"] Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.040274 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zmvcb"] Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.040416 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.135055 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-catalog-content\") pod \"redhat-marketplace-zmvcb\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.135339 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-utilities\") pod \"redhat-marketplace-zmvcb\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.135420 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmdsd\" (UniqueName: \"kubernetes.io/projected/a43ae6f8-5b57-4de9-994c-18169cee9335-kube-api-access-zmdsd\") pod \"redhat-marketplace-zmvcb\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.236828 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmdsd\" (UniqueName: \"kubernetes.io/projected/a43ae6f8-5b57-4de9-994c-18169cee9335-kube-api-access-zmdsd\") pod \"redhat-marketplace-zmvcb\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.236977 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-catalog-content\") pod \"redhat-marketplace-zmvcb\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.237000 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-utilities\") pod \"redhat-marketplace-zmvcb\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.237454 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-utilities\") pod \"redhat-marketplace-zmvcb\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.237960 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-catalog-content\") pod \"redhat-marketplace-zmvcb\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.258290 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmdsd\" (UniqueName: \"kubernetes.io/projected/a43ae6f8-5b57-4de9-994c-18169cee9335-kube-api-access-zmdsd\") pod \"redhat-marketplace-zmvcb\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.360963 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.855942 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zmvcb"] Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.863774 4871 generic.go:334] "Generic (PLEG): container finished" podID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerID="bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96" exitCode=0 Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.863839 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdf5s" event={"ID":"2e10f5ec-3a61-49f6-a7d4-bf540878b04e","Type":"ContainerDied","Data":"bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96"} Oct 08 00:42:47 crc kubenswrapper[4871]: W1008 00:42:47.897779 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda43ae6f8_5b57_4de9_994c_18169cee9335.slice/crio-e7c7c7ef9cafd95fee485ba90bff320d6f092b85c41b3137a4dda814467e4818 WatchSource:0}: Error finding container e7c7c7ef9cafd95fee485ba90bff320d6f092b85c41b3137a4dda814467e4818: Status 404 returned error can't find the container with id e7c7c7ef9cafd95fee485ba90bff320d6f092b85c41b3137a4dda814467e4818 Oct 08 00:42:47 crc kubenswrapper[4871]: I1008 00:42:47.982538 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:42:47 crc kubenswrapper[4871]: E1008 00:42:47.982988 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:42:48 crc kubenswrapper[4871]: I1008 00:42:48.881193 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdf5s" event={"ID":"2e10f5ec-3a61-49f6-a7d4-bf540878b04e","Type":"ContainerStarted","Data":"84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7"} Oct 08 00:42:48 crc kubenswrapper[4871]: I1008 00:42:48.884815 4871 generic.go:334] "Generic (PLEG): container finished" podID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerID="ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063" exitCode=0 Oct 08 00:42:48 crc kubenswrapper[4871]: I1008 00:42:48.884852 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zmvcb" event={"ID":"a43ae6f8-5b57-4de9-994c-18169cee9335","Type":"ContainerDied","Data":"ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063"} Oct 08 00:42:48 crc kubenswrapper[4871]: I1008 00:42:48.884875 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zmvcb" event={"ID":"a43ae6f8-5b57-4de9-994c-18169cee9335","Type":"ContainerStarted","Data":"e7c7c7ef9cafd95fee485ba90bff320d6f092b85c41b3137a4dda814467e4818"} Oct 08 00:42:48 crc kubenswrapper[4871]: I1008 00:42:48.917824 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fdf5s" podStartSLOduration=2.471061485 podStartE2EDuration="4.917803294s" podCreationTimestamp="2025-10-08 00:42:44 +0000 UTC" firstStartedPulling="2025-10-08 00:42:45.839777493 +0000 UTC m=+9239.642475566" lastFinishedPulling="2025-10-08 00:42:48.286519302 +0000 UTC m=+9242.089217375" observedRunningTime="2025-10-08 00:42:48.910287647 +0000 UTC m=+9242.712985760" watchObservedRunningTime="2025-10-08 00:42:48.917803294 +0000 UTC m=+9242.720501377" Oct 08 00:42:50 crc kubenswrapper[4871]: I1008 00:42:50.952078 4871 generic.go:334] "Generic (PLEG): container finished" podID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerID="cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0" exitCode=0 Oct 08 00:42:50 crc kubenswrapper[4871]: I1008 00:42:50.952406 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zmvcb" event={"ID":"a43ae6f8-5b57-4de9-994c-18169cee9335","Type":"ContainerDied","Data":"cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0"} Oct 08 00:42:51 crc kubenswrapper[4871]: I1008 00:42:51.965756 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zmvcb" event={"ID":"a43ae6f8-5b57-4de9-994c-18169cee9335","Type":"ContainerStarted","Data":"eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b"} Oct 08 00:42:51 crc kubenswrapper[4871]: I1008 00:42:51.991974 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zmvcb" podStartSLOduration=3.488077079 podStartE2EDuration="5.991948222s" podCreationTimestamp="2025-10-08 00:42:46 +0000 UTC" firstStartedPulling="2025-10-08 00:42:48.887743077 +0000 UTC m=+9242.690441150" lastFinishedPulling="2025-10-08 00:42:51.39161422 +0000 UTC m=+9245.194312293" observedRunningTime="2025-10-08 00:42:51.985035792 +0000 UTC m=+9245.787733875" watchObservedRunningTime="2025-10-08 00:42:51.991948222 +0000 UTC m=+9245.794646315" Oct 08 00:42:54 crc kubenswrapper[4871]: I1008 00:42:54.946565 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:54 crc kubenswrapper[4871]: I1008 00:42:54.947087 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:55 crc kubenswrapper[4871]: I1008 00:42:55.040099 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:55 crc kubenswrapper[4871]: I1008 00:42:55.104494 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:56 crc kubenswrapper[4871]: I1008 00:42:56.185664 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fdf5s"] Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.026317 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fdf5s" podUID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerName="registry-server" containerID="cri-o://84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7" gracePeriod=2 Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.361670 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.362023 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.424183 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.675113 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.868424 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-catalog-content\") pod \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.868650 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-utilities\") pod \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.868708 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kklc\" (UniqueName: \"kubernetes.io/projected/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-kube-api-access-4kklc\") pod \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\" (UID: \"2e10f5ec-3a61-49f6-a7d4-bf540878b04e\") " Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.869707 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-utilities" (OuterVolumeSpecName: "utilities") pod "2e10f5ec-3a61-49f6-a7d4-bf540878b04e" (UID: "2e10f5ec-3a61-49f6-a7d4-bf540878b04e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.877359 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-kube-api-access-4kklc" (OuterVolumeSpecName: "kube-api-access-4kklc") pod "2e10f5ec-3a61-49f6-a7d4-bf540878b04e" (UID: "2e10f5ec-3a61-49f6-a7d4-bf540878b04e"). InnerVolumeSpecName "kube-api-access-4kklc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.954761 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e10f5ec-3a61-49f6-a7d4-bf540878b04e" (UID: "2e10f5ec-3a61-49f6-a7d4-bf540878b04e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.971399 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.971642 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:42:57 crc kubenswrapper[4871]: I1008 00:42:57.971728 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kklc\" (UniqueName: \"kubernetes.io/projected/2e10f5ec-3a61-49f6-a7d4-bf540878b04e-kube-api-access-4kklc\") on node \"crc\" DevicePath \"\"" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.042446 4871 generic.go:334] "Generic (PLEG): container finished" podID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerID="84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7" exitCode=0 Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.042523 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdf5s" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.042578 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdf5s" event={"ID":"2e10f5ec-3a61-49f6-a7d4-bf540878b04e","Type":"ContainerDied","Data":"84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7"} Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.042634 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdf5s" event={"ID":"2e10f5ec-3a61-49f6-a7d4-bf540878b04e","Type":"ContainerDied","Data":"701f24397b8551babe3a6bb84e97de67e0e14cc910171d1b38d885bee5806f36"} Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.042665 4871 scope.go:117] "RemoveContainer" containerID="84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.088618 4871 scope.go:117] "RemoveContainer" containerID="bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.099061 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fdf5s"] Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.115605 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fdf5s"] Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.124332 4871 scope.go:117] "RemoveContainer" containerID="1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.124420 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.149419 4871 scope.go:117] "RemoveContainer" containerID="84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7" Oct 08 00:42:58 crc kubenswrapper[4871]: E1008 00:42:58.149949 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7\": container with ID starting with 84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7 not found: ID does not exist" containerID="84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.149982 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7"} err="failed to get container status \"84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7\": rpc error: code = NotFound desc = could not find container \"84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7\": container with ID starting with 84c44ee4e2aca5f3f02a92106c2bbfbc0ce6dcc582b8346865a88d3a1d3bfde7 not found: ID does not exist" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.150003 4871 scope.go:117] "RemoveContainer" containerID="bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96" Oct 08 00:42:58 crc kubenswrapper[4871]: E1008 00:42:58.150491 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96\": container with ID starting with bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96 not found: ID does not exist" containerID="bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.150548 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96"} err="failed to get container status \"bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96\": rpc error: code = NotFound desc = could not find container \"bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96\": container with ID starting with bbd4c2ad943f6165988df611d7cdb2a24bfc8f12181bd5c7148a5f4f42549f96 not found: ID does not exist" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.150591 4871 scope.go:117] "RemoveContainer" containerID="1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1" Oct 08 00:42:58 crc kubenswrapper[4871]: E1008 00:42:58.151028 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1\": container with ID starting with 1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1 not found: ID does not exist" containerID="1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.151053 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1"} err="failed to get container status \"1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1\": rpc error: code = NotFound desc = could not find container \"1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1\": container with ID starting with 1c2b5fe426cbc573d601b0a98fccf41838cdd1238f09bcaf8ee8e6948e67f3a1 not found: ID does not exist" Oct 08 00:42:58 crc kubenswrapper[4871]: I1008 00:42:58.997531 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" path="/var/lib/kubelet/pods/2e10f5ec-3a61-49f6-a7d4-bf540878b04e/volumes" Oct 08 00:42:59 crc kubenswrapper[4871]: I1008 00:42:59.379960 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zmvcb"] Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.062161 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zmvcb" podUID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerName="registry-server" containerID="cri-o://eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b" gracePeriod=2 Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.628591 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.739649 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-catalog-content\") pod \"a43ae6f8-5b57-4de9-994c-18169cee9335\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.739837 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmdsd\" (UniqueName: \"kubernetes.io/projected/a43ae6f8-5b57-4de9-994c-18169cee9335-kube-api-access-zmdsd\") pod \"a43ae6f8-5b57-4de9-994c-18169cee9335\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.739904 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-utilities\") pod \"a43ae6f8-5b57-4de9-994c-18169cee9335\" (UID: \"a43ae6f8-5b57-4de9-994c-18169cee9335\") " Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.741010 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-utilities" (OuterVolumeSpecName: "utilities") pod "a43ae6f8-5b57-4de9-994c-18169cee9335" (UID: "a43ae6f8-5b57-4de9-994c-18169cee9335"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.745026 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a43ae6f8-5b57-4de9-994c-18169cee9335-kube-api-access-zmdsd" (OuterVolumeSpecName: "kube-api-access-zmdsd") pod "a43ae6f8-5b57-4de9-994c-18169cee9335" (UID: "a43ae6f8-5b57-4de9-994c-18169cee9335"). InnerVolumeSpecName "kube-api-access-zmdsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.757127 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a43ae6f8-5b57-4de9-994c-18169cee9335" (UID: "a43ae6f8-5b57-4de9-994c-18169cee9335"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.842725 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.842770 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmdsd\" (UniqueName: \"kubernetes.io/projected/a43ae6f8-5b57-4de9-994c-18169cee9335-kube-api-access-zmdsd\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.842785 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a43ae6f8-5b57-4de9-994c-18169cee9335-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:00 crc kubenswrapper[4871]: I1008 00:43:00.984635 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:43:00 crc kubenswrapper[4871]: E1008 00:43:00.985055 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.097100 4871 generic.go:334] "Generic (PLEG): container finished" podID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerID="eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b" exitCode=0 Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.097150 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zmvcb" event={"ID":"a43ae6f8-5b57-4de9-994c-18169cee9335","Type":"ContainerDied","Data":"eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b"} Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.097180 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zmvcb" event={"ID":"a43ae6f8-5b57-4de9-994c-18169cee9335","Type":"ContainerDied","Data":"e7c7c7ef9cafd95fee485ba90bff320d6f092b85c41b3137a4dda814467e4818"} Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.097197 4871 scope.go:117] "RemoveContainer" containerID="eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.097231 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zmvcb" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.128497 4871 scope.go:117] "RemoveContainer" containerID="cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.139477 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zmvcb"] Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.149447 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zmvcb"] Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.164452 4871 scope.go:117] "RemoveContainer" containerID="ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.227681 4871 scope.go:117] "RemoveContainer" containerID="eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b" Oct 08 00:43:01 crc kubenswrapper[4871]: E1008 00:43:01.228328 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b\": container with ID starting with eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b not found: ID does not exist" containerID="eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.228372 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b"} err="failed to get container status \"eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b\": rpc error: code = NotFound desc = could not find container \"eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b\": container with ID starting with eb41d4cc4fd97b5e885516b22e943b7288ba4fee4f199d72247f08a31bce811b not found: ID does not exist" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.228403 4871 scope.go:117] "RemoveContainer" containerID="cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0" Oct 08 00:43:01 crc kubenswrapper[4871]: E1008 00:43:01.228917 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0\": container with ID starting with cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0 not found: ID does not exist" containerID="cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.229013 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0"} err="failed to get container status \"cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0\": rpc error: code = NotFound desc = could not find container \"cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0\": container with ID starting with cf62292d3f7d22603240f5767f31ebe00e8d1f44d2f2a99ee3adc7e15b41f6c0 not found: ID does not exist" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.229085 4871 scope.go:117] "RemoveContainer" containerID="ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063" Oct 08 00:43:01 crc kubenswrapper[4871]: E1008 00:43:01.229671 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063\": container with ID starting with ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063 not found: ID does not exist" containerID="ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063" Oct 08 00:43:01 crc kubenswrapper[4871]: I1008 00:43:01.229718 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063"} err="failed to get container status \"ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063\": rpc error: code = NotFound desc = could not find container \"ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063\": container with ID starting with ce92a497b353e01a8d7a3f1eb493fbb0b6e8e098a1156f7529245171fb14c063 not found: ID does not exist" Oct 08 00:43:02 crc kubenswrapper[4871]: I1008 00:43:02.117940 4871 generic.go:334] "Generic (PLEG): container finished" podID="ac814cee-d737-4c1f-b486-93d99024c197" containerID="167c976dd15143d9cb0f182cc23579aaad3e4f8aa514f7a579b71a7904b595fc" exitCode=0 Oct 08 00:43:02 crc kubenswrapper[4871]: I1008 00:43:02.117993 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" event={"ID":"ac814cee-d737-4c1f-b486-93d99024c197","Type":"ContainerDied","Data":"167c976dd15143d9cb0f182cc23579aaad3e4f8aa514f7a579b71a7904b595fc"} Oct 08 00:43:02 crc kubenswrapper[4871]: I1008 00:43:02.997354 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a43ae6f8-5b57-4de9-994c-18169cee9335" path="/var/lib/kubelet/pods/a43ae6f8-5b57-4de9-994c-18169cee9335/volumes" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.701932 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.817762 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ssh-key\") pod \"ac814cee-d737-4c1f-b486-93d99024c197\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.818023 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ceph\") pod \"ac814cee-d737-4c1f-b486-93d99024c197\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.818080 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brxps\" (UniqueName: \"kubernetes.io/projected/ac814cee-d737-4c1f-b486-93d99024c197-kube-api-access-brxps\") pod \"ac814cee-d737-4c1f-b486-93d99024c197\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.818178 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-agent-neutron-config-0\") pod \"ac814cee-d737-4c1f-b486-93d99024c197\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.818271 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-combined-ca-bundle\") pod \"ac814cee-d737-4c1f-b486-93d99024c197\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.818329 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-inventory\") pod \"ac814cee-d737-4c1f-b486-93d99024c197\" (UID: \"ac814cee-d737-4c1f-b486-93d99024c197\") " Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.824303 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac814cee-d737-4c1f-b486-93d99024c197-kube-api-access-brxps" (OuterVolumeSpecName: "kube-api-access-brxps") pod "ac814cee-d737-4c1f-b486-93d99024c197" (UID: "ac814cee-d737-4c1f-b486-93d99024c197"). InnerVolumeSpecName "kube-api-access-brxps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.830023 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "ac814cee-d737-4c1f-b486-93d99024c197" (UID: "ac814cee-d737-4c1f-b486-93d99024c197"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.843538 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ceph" (OuterVolumeSpecName: "ceph") pod "ac814cee-d737-4c1f-b486-93d99024c197" (UID: "ac814cee-d737-4c1f-b486-93d99024c197"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.885411 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ac814cee-d737-4c1f-b486-93d99024c197" (UID: "ac814cee-d737-4c1f-b486-93d99024c197"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.885964 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-inventory" (OuterVolumeSpecName: "inventory") pod "ac814cee-d737-4c1f-b486-93d99024c197" (UID: "ac814cee-d737-4c1f-b486-93d99024c197"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.888884 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "ac814cee-d737-4c1f-b486-93d99024c197" (UID: "ac814cee-d737-4c1f-b486-93d99024c197"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.922326 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brxps\" (UniqueName: \"kubernetes.io/projected/ac814cee-d737-4c1f-b486-93d99024c197-kube-api-access-brxps\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.922608 4871 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.922665 4871 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.922686 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.922705 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:03 crc kubenswrapper[4871]: I1008 00:43:03.922751 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac814cee-d737-4c1f-b486-93d99024c197-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:04 crc kubenswrapper[4871]: I1008 00:43:04.140027 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" event={"ID":"ac814cee-d737-4c1f-b486-93d99024c197","Type":"ContainerDied","Data":"dc448799a0acb23d80a19d7603db8a6d698d68c010f1911b53a650b2978bd35a"} Oct 08 00:43:04 crc kubenswrapper[4871]: I1008 00:43:04.140272 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc448799a0acb23d80a19d7603db8a6d698d68c010f1911b53a650b2978bd35a" Oct 08 00:43:04 crc kubenswrapper[4871]: I1008 00:43:04.140115 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-rfmrs" Oct 08 00:43:14 crc kubenswrapper[4871]: I1008 00:43:14.983009 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:43:14 crc kubenswrapper[4871]: E1008 00:43:14.984167 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.152698 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.153378 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="426135f0-8fe9-40e8-9a1d-f95c000de11a" containerName="nova-cell0-conductor-conductor" containerID="cri-o://463df52f01d8665c2d960be5d6a0bafa81b877c3ce9fde41cf2bbd3ab37dd67b" gracePeriod=30 Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.944387 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.944613 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="6e94c8d0-f89e-4215-8c9c-158cfead9a4a" containerName="nova-cell1-conductor-conductor" containerID="cri-o://6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221" gracePeriod=30 Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.973223 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk"] Oct 08 00:43:25 crc kubenswrapper[4871]: E1008 00:43:25.973741 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerName="extract-utilities" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.973759 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerName="extract-utilities" Oct 08 00:43:25 crc kubenswrapper[4871]: E1008 00:43:25.973776 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerName="registry-server" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.973784 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerName="registry-server" Oct 08 00:43:25 crc kubenswrapper[4871]: E1008 00:43:25.973808 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerName="extract-content" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.973814 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerName="extract-content" Oct 08 00:43:25 crc kubenswrapper[4871]: E1008 00:43:25.973840 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerName="extract-content" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.973846 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerName="extract-content" Oct 08 00:43:25 crc kubenswrapper[4871]: E1008 00:43:25.973857 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerName="extract-utilities" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.973864 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerName="extract-utilities" Oct 08 00:43:25 crc kubenswrapper[4871]: E1008 00:43:25.973893 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac814cee-d737-4c1f-b486-93d99024c197" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.973902 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac814cee-d737-4c1f-b486-93d99024c197" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 08 00:43:25 crc kubenswrapper[4871]: E1008 00:43:25.973913 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerName="registry-server" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.973918 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerName="registry-server" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.974102 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a43ae6f8-5b57-4de9-994c-18169cee9335" containerName="registry-server" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.974116 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac814cee-d737-4c1f-b486-93d99024c197" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.974135 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e10f5ec-3a61-49f6-a7d4-bf540878b04e" containerName="registry-server" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.974930 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.976732 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.978197 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.978284 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-ljqpk" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.978405 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.978974 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.979093 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.979146 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 00:43:25 crc kubenswrapper[4871]: I1008 00:43:25.996190 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk"] Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.061349 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.061396 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.061930 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.061956 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.062022 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.062039 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68dzn\" (UniqueName: \"kubernetes.io/projected/b8b25c96-823b-433c-8512-cf364233ebe0-kube-api-access-68dzn\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.062084 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.062122 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.062139 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.062181 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.062200 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.150318 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.151036 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a360e652-467f-4d96-9232-d944191ee751" containerName="nova-scheduler-scheduler" containerID="cri-o://dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" gracePeriod=30 Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.160700 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.160937 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-log" containerID="cri-o://36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1" gracePeriod=30 Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.161108 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-api" containerID="cri-o://ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1" gracePeriod=30 Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.163879 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.163928 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68dzn\" (UniqueName: \"kubernetes.io/projected/b8b25c96-823b-433c-8512-cf364233ebe0-kube-api-access-68dzn\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.163982 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.164027 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.164057 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.164127 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.164162 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.164215 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.164251 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.164540 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.164571 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.166300 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.166638 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.171546 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.172267 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.174618 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.185277 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.185419 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.185883 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.196018 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.201052 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.212598 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68dzn\" (UniqueName: \"kubernetes.io/projected/b8b25c96-823b-433c-8512-cf364233ebe0-kube-api-access-68dzn\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.214508 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.214775 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-log" containerID="cri-o://6a75584a8f6ac34bebf2de3afad9e86ebd85f9ba6fb3d711a0941a19bd274f39" gracePeriod=30 Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.214951 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-metadata" containerID="cri-o://2a76e013141cc21f95e127a57c2b86803526962045fe5a12a9ad4f1f8656b434" gracePeriod=30 Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.317383 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.436347 4871 generic.go:334] "Generic (PLEG): container finished" podID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerID="36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1" exitCode=143 Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.436563 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a313c68a-aeef-4003-a77d-97fb7ba6b22c","Type":"ContainerDied","Data":"36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1"} Oct 08 00:43:26 crc kubenswrapper[4871]: E1008 00:43:26.444946 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 00:43:26 crc kubenswrapper[4871]: E1008 00:43:26.446974 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 00:43:26 crc kubenswrapper[4871]: E1008 00:43:26.450591 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 00:43:26 crc kubenswrapper[4871]: E1008 00:43:26.450645 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a360e652-467f-4d96-9232-d944191ee751" containerName="nova-scheduler-scheduler" Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.452471 4871 generic.go:334] "Generic (PLEG): container finished" podID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerID="6a75584a8f6ac34bebf2de3afad9e86ebd85f9ba6fb3d711a0941a19bd274f39" exitCode=143 Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.452507 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fe15a5e-9ee6-47ba-a450-044b62740488","Type":"ContainerDied","Data":"6a75584a8f6ac34bebf2de3afad9e86ebd85f9ba6fb3d711a0941a19bd274f39"} Oct 08 00:43:26 crc kubenswrapper[4871]: I1008 00:43:26.993995 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk"] Oct 08 00:43:27 crc kubenswrapper[4871]: I1008 00:43:27.466280 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" event={"ID":"b8b25c96-823b-433c-8512-cf364233ebe0","Type":"ContainerStarted","Data":"9f5217d4231849819da4780647f55450fad380b75fb745744ac4d6fd621aa466"} Oct 08 00:43:27 crc kubenswrapper[4871]: E1008 00:43:27.467993 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 00:43:27 crc kubenswrapper[4871]: E1008 00:43:27.471107 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 00:43:27 crc kubenswrapper[4871]: E1008 00:43:27.473112 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 00:43:27 crc kubenswrapper[4871]: E1008 00:43:27.473170 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="6e94c8d0-f89e-4215-8c9c-158cfead9a4a" containerName="nova-cell1-conductor-conductor" Oct 08 00:43:28 crc kubenswrapper[4871]: E1008 00:43:28.310711 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="463df52f01d8665c2d960be5d6a0bafa81b877c3ce9fde41cf2bbd3ab37dd67b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 00:43:28 crc kubenswrapper[4871]: E1008 00:43:28.312911 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="463df52f01d8665c2d960be5d6a0bafa81b877c3ce9fde41cf2bbd3ab37dd67b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 00:43:28 crc kubenswrapper[4871]: E1008 00:43:28.318479 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="463df52f01d8665c2d960be5d6a0bafa81b877c3ce9fde41cf2bbd3ab37dd67b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 00:43:28 crc kubenswrapper[4871]: E1008 00:43:28.318536 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="426135f0-8fe9-40e8-9a1d-f95c000de11a" containerName="nova-cell0-conductor-conductor" Oct 08 00:43:28 crc kubenswrapper[4871]: I1008 00:43:28.477187 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" event={"ID":"b8b25c96-823b-433c-8512-cf364233ebe0","Type":"ContainerStarted","Data":"28948d188d5b1fdd0e9987a6cd018f729af05f8c6ecdfdbbaf9cb4a2cc25bbbe"} Oct 08 00:43:28 crc kubenswrapper[4871]: I1008 00:43:28.510937 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" podStartSLOduration=2.999679876 podStartE2EDuration="3.510908646s" podCreationTimestamp="2025-10-08 00:43:25 +0000 UTC" firstStartedPulling="2025-10-08 00:43:26.996541702 +0000 UTC m=+9280.799239775" lastFinishedPulling="2025-10-08 00:43:27.507770432 +0000 UTC m=+9281.310468545" observedRunningTime="2025-10-08 00:43:28.494772524 +0000 UTC m=+9282.297470597" watchObservedRunningTime="2025-10-08 00:43:28.510908646 +0000 UTC m=+9282.313606759" Oct 08 00:43:28 crc kubenswrapper[4871]: I1008 00:43:28.982951 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:43:28 crc kubenswrapper[4871]: E1008 00:43:28.983350 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:43:29 crc kubenswrapper[4871]: I1008 00:43:29.373018 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": read tcp 10.217.0.2:59546->10.217.1.81:8775: read: connection reset by peer" Oct 08 00:43:29 crc kubenswrapper[4871]: I1008 00:43:29.373048 4871 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": read tcp 10.217.0.2:59556->10.217.1.81:8775: read: connection reset by peer" Oct 08 00:43:29 crc kubenswrapper[4871]: I1008 00:43:29.502998 4871 generic.go:334] "Generic (PLEG): container finished" podID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerID="2a76e013141cc21f95e127a57c2b86803526962045fe5a12a9ad4f1f8656b434" exitCode=0 Oct 08 00:43:29 crc kubenswrapper[4871]: I1008 00:43:29.503067 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fe15a5e-9ee6-47ba-a450-044b62740488","Type":"ContainerDied","Data":"2a76e013141cc21f95e127a57c2b86803526962045fe5a12a9ad4f1f8656b434"} Oct 08 00:43:29 crc kubenswrapper[4871]: I1008 00:43:29.953868 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.062962 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe15a5e-9ee6-47ba-a450-044b62740488-logs\") pod \"7fe15a5e-9ee6-47ba-a450-044b62740488\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.063117 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-config-data\") pod \"7fe15a5e-9ee6-47ba-a450-044b62740488\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.063169 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d2x8\" (UniqueName: \"kubernetes.io/projected/7fe15a5e-9ee6-47ba-a450-044b62740488-kube-api-access-7d2x8\") pod \"7fe15a5e-9ee6-47ba-a450-044b62740488\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.063194 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-combined-ca-bundle\") pod \"7fe15a5e-9ee6-47ba-a450-044b62740488\" (UID: \"7fe15a5e-9ee6-47ba-a450-044b62740488\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.064183 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fe15a5e-9ee6-47ba-a450-044b62740488-logs" (OuterVolumeSpecName: "logs") pod "7fe15a5e-9ee6-47ba-a450-044b62740488" (UID: "7fe15a5e-9ee6-47ba-a450-044b62740488"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.074250 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe15a5e-9ee6-47ba-a450-044b62740488-kube-api-access-7d2x8" (OuterVolumeSpecName: "kube-api-access-7d2x8") pod "7fe15a5e-9ee6-47ba-a450-044b62740488" (UID: "7fe15a5e-9ee6-47ba-a450-044b62740488"). InnerVolumeSpecName "kube-api-access-7d2x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.132939 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fe15a5e-9ee6-47ba-a450-044b62740488" (UID: "7fe15a5e-9ee6-47ba-a450-044b62740488"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.159950 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-config-data" (OuterVolumeSpecName: "config-data") pod "7fe15a5e-9ee6-47ba-a450-044b62740488" (UID: "7fe15a5e-9ee6-47ba-a450-044b62740488"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.163675 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.166167 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.166187 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d2x8\" (UniqueName: \"kubernetes.io/projected/7fe15a5e-9ee6-47ba-a450-044b62740488-kube-api-access-7d2x8\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.166198 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe15a5e-9ee6-47ba-a450-044b62740488-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.166209 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fe15a5e-9ee6-47ba-a450-044b62740488-logs\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.267235 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-combined-ca-bundle\") pod \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.267749 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82n4l\" (UniqueName: \"kubernetes.io/projected/a313c68a-aeef-4003-a77d-97fb7ba6b22c-kube-api-access-82n4l\") pod \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.267971 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-config-data\") pod \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.268006 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a313c68a-aeef-4003-a77d-97fb7ba6b22c-logs\") pod \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\" (UID: \"a313c68a-aeef-4003-a77d-97fb7ba6b22c\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.268603 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a313c68a-aeef-4003-a77d-97fb7ba6b22c-logs" (OuterVolumeSpecName: "logs") pod "a313c68a-aeef-4003-a77d-97fb7ba6b22c" (UID: "a313c68a-aeef-4003-a77d-97fb7ba6b22c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.268832 4871 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a313c68a-aeef-4003-a77d-97fb7ba6b22c-logs\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.277489 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a313c68a-aeef-4003-a77d-97fb7ba6b22c-kube-api-access-82n4l" (OuterVolumeSpecName: "kube-api-access-82n4l") pod "a313c68a-aeef-4003-a77d-97fb7ba6b22c" (UID: "a313c68a-aeef-4003-a77d-97fb7ba6b22c"). InnerVolumeSpecName "kube-api-access-82n4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.307322 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-config-data" (OuterVolumeSpecName: "config-data") pod "a313c68a-aeef-4003-a77d-97fb7ba6b22c" (UID: "a313c68a-aeef-4003-a77d-97fb7ba6b22c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.314088 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a313c68a-aeef-4003-a77d-97fb7ba6b22c" (UID: "a313c68a-aeef-4003-a77d-97fb7ba6b22c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.371063 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.371106 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82n4l\" (UniqueName: \"kubernetes.io/projected/a313c68a-aeef-4003-a77d-97fb7ba6b22c-kube-api-access-82n4l\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.371124 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a313c68a-aeef-4003-a77d-97fb7ba6b22c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.414782 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.472233 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n72pd\" (UniqueName: \"kubernetes.io/projected/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-kube-api-access-n72pd\") pod \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.472416 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-combined-ca-bundle\") pod \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.472522 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-config-data\") pod \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\" (UID: \"6e94c8d0-f89e-4215-8c9c-158cfead9a4a\") " Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.478374 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-kube-api-access-n72pd" (OuterVolumeSpecName: "kube-api-access-n72pd") pod "6e94c8d0-f89e-4215-8c9c-158cfead9a4a" (UID: "6e94c8d0-f89e-4215-8c9c-158cfead9a4a"). InnerVolumeSpecName "kube-api-access-n72pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.500353 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-config-data" (OuterVolumeSpecName: "config-data") pod "6e94c8d0-f89e-4215-8c9c-158cfead9a4a" (UID: "6e94c8d0-f89e-4215-8c9c-158cfead9a4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.501197 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e94c8d0-f89e-4215-8c9c-158cfead9a4a" (UID: "6e94c8d0-f89e-4215-8c9c-158cfead9a4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.516731 4871 generic.go:334] "Generic (PLEG): container finished" podID="6e94c8d0-f89e-4215-8c9c-158cfead9a4a" containerID="6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221" exitCode=0 Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.516785 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6e94c8d0-f89e-4215-8c9c-158cfead9a4a","Type":"ContainerDied","Data":"6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221"} Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.516858 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.516871 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6e94c8d0-f89e-4215-8c9c-158cfead9a4a","Type":"ContainerDied","Data":"fc1d17ff6504d420075fc6b882eb1540d2c6cb6d01e78ed391f9c6dad673d2e8"} Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.516893 4871 scope.go:117] "RemoveContainer" containerID="6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.520386 4871 generic.go:334] "Generic (PLEG): container finished" podID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerID="ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1" exitCode=0 Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.520439 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a313c68a-aeef-4003-a77d-97fb7ba6b22c","Type":"ContainerDied","Data":"ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1"} Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.520463 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a313c68a-aeef-4003-a77d-97fb7ba6b22c","Type":"ContainerDied","Data":"34ab9a938bd7ce476e94621b1fe895d0f1bb56afc324e758fdb676767eb61c77"} Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.520518 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.527341 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7fe15a5e-9ee6-47ba-a450-044b62740488","Type":"ContainerDied","Data":"7f92d5aaefeada9eb78afef7f6706ddb9c30f3bd598b53d70bfdf5ea6256cd85"} Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.527407 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.582139 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.582179 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n72pd\" (UniqueName: \"kubernetes.io/projected/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-kube-api-access-n72pd\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.582190 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e94c8d0-f89e-4215-8c9c-158cfead9a4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.601524 4871 scope.go:117] "RemoveContainer" containerID="6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221" Oct 08 00:43:30 crc kubenswrapper[4871]: E1008 00:43:30.607930 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221\": container with ID starting with 6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221 not found: ID does not exist" containerID="6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.608001 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221"} err="failed to get container status \"6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221\": rpc error: code = NotFound desc = could not find container \"6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221\": container with ID starting with 6b77b428218eb8d028e0531d74bad518ed200f5fea5d8dd82ec85294a187c221 not found: ID does not exist" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.608076 4871 scope.go:117] "RemoveContainer" containerID="ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.615554 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.642046 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.665651 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.686528 4871 scope.go:117] "RemoveContainer" containerID="36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.707220 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: E1008 00:43:30.707716 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e94c8d0-f89e-4215-8c9c-158cfead9a4a" containerName="nova-cell1-conductor-conductor" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.707734 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e94c8d0-f89e-4215-8c9c-158cfead9a4a" containerName="nova-cell1-conductor-conductor" Oct 08 00:43:30 crc kubenswrapper[4871]: E1008 00:43:30.707761 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-log" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.707769 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-log" Oct 08 00:43:30 crc kubenswrapper[4871]: E1008 00:43:30.707782 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-metadata" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.707807 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-metadata" Oct 08 00:43:30 crc kubenswrapper[4871]: E1008 00:43:30.707843 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-api" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.707851 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-api" Oct 08 00:43:30 crc kubenswrapper[4871]: E1008 00:43:30.707874 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-log" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.707882 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-log" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.708115 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-log" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.708136 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-api" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.708151 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" containerName="nova-api-log" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.708168 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" containerName="nova-metadata-metadata" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.708181 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e94c8d0-f89e-4215-8c9c-158cfead9a4a" containerName="nova-cell1-conductor-conductor" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.709548 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.709742 4871 scope.go:117] "RemoveContainer" containerID="ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1" Oct 08 00:43:30 crc kubenswrapper[4871]: E1008 00:43:30.710210 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1\": container with ID starting with ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1 not found: ID does not exist" containerID="ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.710236 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1"} err="failed to get container status \"ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1\": rpc error: code = NotFound desc = could not find container \"ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1\": container with ID starting with ccbf18a1c208df641b2eba1fca424786cb9d0dd92fe0ed6c2294f75d1942d7f1 not found: ID does not exist" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.710256 4871 scope.go:117] "RemoveContainer" containerID="36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1" Oct 08 00:43:30 crc kubenswrapper[4871]: E1008 00:43:30.710502 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1\": container with ID starting with 36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1 not found: ID does not exist" containerID="36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.710537 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1"} err="failed to get container status \"36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1\": rpc error: code = NotFound desc = could not find container \"36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1\": container with ID starting with 36b6ea3c178f801a89cee2ff218d176649b2a0ae6a61a4a650154a9c0b08d8b1 not found: ID does not exist" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.710571 4871 scope.go:117] "RemoveContainer" containerID="2a76e013141cc21f95e127a57c2b86803526962045fe5a12a9ad4f1f8656b434" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.711644 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.733595 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.746970 4871 scope.go:117] "RemoveContainer" containerID="6a75584a8f6ac34bebf2de3afad9e86ebd85f9ba6fb3d711a0941a19bd274f39" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.754943 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.764877 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.782283 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.794051 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f07d2a-d7a7-46b1-9e36-faff79d08add-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.794140 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9f07d2a-d7a7-46b1-9e36-faff79d08add-config-data\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.794207 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm8w2\" (UniqueName: \"kubernetes.io/projected/d9f07d2a-d7a7-46b1-9e36-faff79d08add-kube-api-access-wm8w2\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.794299 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9f07d2a-d7a7-46b1-9e36-faff79d08add-logs\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.794398 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.796264 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.799031 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.805391 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.807266 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.814769 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.817608 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.827818 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.895942 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm8w2\" (UniqueName: \"kubernetes.io/projected/d9f07d2a-d7a7-46b1-9e36-faff79d08add-kube-api-access-wm8w2\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.895985 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e277fae-f0f9-469b-87da-b7c88f40871f-logs\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896039 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9f07d2a-d7a7-46b1-9e36-faff79d08add-logs\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896071 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57cb8659-97fb-41d1-b2cf-044b95643dd2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"57cb8659-97fb-41d1-b2cf-044b95643dd2\") " pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896095 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm7zj\" (UniqueName: \"kubernetes.io/projected/4e277fae-f0f9-469b-87da-b7c88f40871f-kube-api-access-xm7zj\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896115 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e277fae-f0f9-469b-87da-b7c88f40871f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896206 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e277fae-f0f9-469b-87da-b7c88f40871f-config-data\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896220 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57cb8659-97fb-41d1-b2cf-044b95643dd2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"57cb8659-97fb-41d1-b2cf-044b95643dd2\") " pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896241 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f07d2a-d7a7-46b1-9e36-faff79d08add-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896268 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvkbl\" (UniqueName: \"kubernetes.io/projected/57cb8659-97fb-41d1-b2cf-044b95643dd2-kube-api-access-tvkbl\") pod \"nova-cell1-conductor-0\" (UID: \"57cb8659-97fb-41d1-b2cf-044b95643dd2\") " pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896298 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9f07d2a-d7a7-46b1-9e36-faff79d08add-config-data\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.896627 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9f07d2a-d7a7-46b1-9e36-faff79d08add-logs\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.902234 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9f07d2a-d7a7-46b1-9e36-faff79d08add-config-data\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.904341 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f07d2a-d7a7-46b1-9e36-faff79d08add-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.914353 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm8w2\" (UniqueName: \"kubernetes.io/projected/d9f07d2a-d7a7-46b1-9e36-faff79d08add-kube-api-access-wm8w2\") pod \"nova-api-0\" (UID: \"d9f07d2a-d7a7-46b1-9e36-faff79d08add\") " pod="openstack/nova-api-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.997587 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e277fae-f0f9-469b-87da-b7c88f40871f-logs\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.997899 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57cb8659-97fb-41d1-b2cf-044b95643dd2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"57cb8659-97fb-41d1-b2cf-044b95643dd2\") " pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.997932 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm7zj\" (UniqueName: \"kubernetes.io/projected/4e277fae-f0f9-469b-87da-b7c88f40871f-kube-api-access-xm7zj\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.997951 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e277fae-f0f9-469b-87da-b7c88f40871f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.998055 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e277fae-f0f9-469b-87da-b7c88f40871f-config-data\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.998074 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57cb8659-97fb-41d1-b2cf-044b95643dd2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"57cb8659-97fb-41d1-b2cf-044b95643dd2\") " pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.998105 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvkbl\" (UniqueName: \"kubernetes.io/projected/57cb8659-97fb-41d1-b2cf-044b95643dd2-kube-api-access-tvkbl\") pod \"nova-cell1-conductor-0\" (UID: \"57cb8659-97fb-41d1-b2cf-044b95643dd2\") " pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:30 crc kubenswrapper[4871]: I1008 00:43:30.998107 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e277fae-f0f9-469b-87da-b7c88f40871f-logs\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.002178 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e277fae-f0f9-469b-87da-b7c88f40871f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.003546 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e277fae-f0f9-469b-87da-b7c88f40871f-config-data\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.004766 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57cb8659-97fb-41d1-b2cf-044b95643dd2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"57cb8659-97fb-41d1-b2cf-044b95643dd2\") " pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.009643 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57cb8659-97fb-41d1-b2cf-044b95643dd2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"57cb8659-97fb-41d1-b2cf-044b95643dd2\") " pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.009544 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e94c8d0-f89e-4215-8c9c-158cfead9a4a" path="/var/lib/kubelet/pods/6e94c8d0-f89e-4215-8c9c-158cfead9a4a/volumes" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.011193 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fe15a5e-9ee6-47ba-a450-044b62740488" path="/var/lib/kubelet/pods/7fe15a5e-9ee6-47ba-a450-044b62740488/volumes" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.011849 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a313c68a-aeef-4003-a77d-97fb7ba6b22c" path="/var/lib/kubelet/pods/a313c68a-aeef-4003-a77d-97fb7ba6b22c/volumes" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.013236 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvkbl\" (UniqueName: \"kubernetes.io/projected/57cb8659-97fb-41d1-b2cf-044b95643dd2-kube-api-access-tvkbl\") pod \"nova-cell1-conductor-0\" (UID: \"57cb8659-97fb-41d1-b2cf-044b95643dd2\") " pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.014229 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm7zj\" (UniqueName: \"kubernetes.io/projected/4e277fae-f0f9-469b-87da-b7c88f40871f-kube-api-access-xm7zj\") pod \"nova-metadata-0\" (UID: \"4e277fae-f0f9-469b-87da-b7c88f40871f\") " pod="openstack/nova-metadata-0" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.034382 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.127712 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.136975 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:31 crc kubenswrapper[4871]: E1008 00:43:31.423811 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 00:43:31 crc kubenswrapper[4871]: E1008 00:43:31.425212 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 00:43:31 crc kubenswrapper[4871]: E1008 00:43:31.426476 4871 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 00:43:31 crc kubenswrapper[4871]: E1008 00:43:31.426541 4871 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a360e652-467f-4d96-9232-d944191ee751" containerName="nova-scheduler-scheduler" Oct 08 00:43:31 crc kubenswrapper[4871]: W1008 00:43:31.508705 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9f07d2a_d7a7_46b1_9e36_faff79d08add.slice/crio-3f1987311fc4895bfe9a4e2b6ebf7e96d3b0cf0f5860cfb7ad50877ddd2ef302 WatchSource:0}: Error finding container 3f1987311fc4895bfe9a4e2b6ebf7e96d3b0cf0f5860cfb7ad50877ddd2ef302: Status 404 returned error can't find the container with id 3f1987311fc4895bfe9a4e2b6ebf7e96d3b0cf0f5860cfb7ad50877ddd2ef302 Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.514147 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.557232 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9f07d2a-d7a7-46b1-9e36-faff79d08add","Type":"ContainerStarted","Data":"3f1987311fc4895bfe9a4e2b6ebf7e96d3b0cf0f5860cfb7ad50877ddd2ef302"} Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.658378 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 00:43:31 crc kubenswrapper[4871]: I1008 00:43:31.669257 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.579300 4871 generic.go:334] "Generic (PLEG): container finished" podID="426135f0-8fe9-40e8-9a1d-f95c000de11a" containerID="463df52f01d8665c2d960be5d6a0bafa81b877c3ce9fde41cf2bbd3ab37dd67b" exitCode=0 Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.579842 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"426135f0-8fe9-40e8-9a1d-f95c000de11a","Type":"ContainerDied","Data":"463df52f01d8665c2d960be5d6a0bafa81b877c3ce9fde41cf2bbd3ab37dd67b"} Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.587703 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"57cb8659-97fb-41d1-b2cf-044b95643dd2","Type":"ContainerStarted","Data":"e8128bbd49813d24ca1d205087eaec863158e334b879ee32322697c09737d164"} Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.587743 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"57cb8659-97fb-41d1-b2cf-044b95643dd2","Type":"ContainerStarted","Data":"9fc09bc65364585d3320aa9e01de7543ffb35f9a55f9782aaffb1077b1912d39"} Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.588917 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.599569 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9f07d2a-d7a7-46b1-9e36-faff79d08add","Type":"ContainerStarted","Data":"10e7d6aa57de5641731f1b697765be140e01820ebbe503edbb5aa6047e8bf64c"} Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.599609 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9f07d2a-d7a7-46b1-9e36-faff79d08add","Type":"ContainerStarted","Data":"ea7bb605dbf438850deaca2b7987b2b1c78d4cb8c84f4d4648fe1f9b57b5abac"} Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.625453 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e277fae-f0f9-469b-87da-b7c88f40871f","Type":"ContainerStarted","Data":"611d78a8fc0bc086eb6f8cab11f27f531bd5fb59d5c117ab6610f693c033e4c5"} Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.625496 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e277fae-f0f9-469b-87da-b7c88f40871f","Type":"ContainerStarted","Data":"3bd98e0e303228f72149a367143fbbf81c808588510a44d47c14fb203597e8ef"} Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.625505 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e277fae-f0f9-469b-87da-b7c88f40871f","Type":"ContainerStarted","Data":"59f57fa8b68ba70da8962640e2d1e8574108cca5d24b5d95c2c86ebb9d071bd5"} Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.635686 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.635665322 podStartE2EDuration="2.635665322s" podCreationTimestamp="2025-10-08 00:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:43:32.601882537 +0000 UTC m=+9286.404580610" watchObservedRunningTime="2025-10-08 00:43:32.635665322 +0000 UTC m=+9286.438363395" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.652092 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.652073321 podStartE2EDuration="2.652073321s" podCreationTimestamp="2025-10-08 00:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:43:32.623733809 +0000 UTC m=+9286.426431882" watchObservedRunningTime="2025-10-08 00:43:32.652073321 +0000 UTC m=+9286.454771394" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.666281 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.666256452 podStartE2EDuration="2.666256452s" podCreationTimestamp="2025-10-08 00:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:43:32.652892783 +0000 UTC m=+9286.455590856" watchObservedRunningTime="2025-10-08 00:43:32.666256452 +0000 UTC m=+9286.468954525" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.772856 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.862411 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-combined-ca-bundle\") pod \"426135f0-8fe9-40e8-9a1d-f95c000de11a\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.862544 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-config-data\") pod \"426135f0-8fe9-40e8-9a1d-f95c000de11a\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.862974 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r92x7\" (UniqueName: \"kubernetes.io/projected/426135f0-8fe9-40e8-9a1d-f95c000de11a-kube-api-access-r92x7\") pod \"426135f0-8fe9-40e8-9a1d-f95c000de11a\" (UID: \"426135f0-8fe9-40e8-9a1d-f95c000de11a\") " Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.871187 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/426135f0-8fe9-40e8-9a1d-f95c000de11a-kube-api-access-r92x7" (OuterVolumeSpecName: "kube-api-access-r92x7") pod "426135f0-8fe9-40e8-9a1d-f95c000de11a" (UID: "426135f0-8fe9-40e8-9a1d-f95c000de11a"). InnerVolumeSpecName "kube-api-access-r92x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.900624 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "426135f0-8fe9-40e8-9a1d-f95c000de11a" (UID: "426135f0-8fe9-40e8-9a1d-f95c000de11a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.902952 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-config-data" (OuterVolumeSpecName: "config-data") pod "426135f0-8fe9-40e8-9a1d-f95c000de11a" (UID: "426135f0-8fe9-40e8-9a1d-f95c000de11a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.966097 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r92x7\" (UniqueName: \"kubernetes.io/projected/426135f0-8fe9-40e8-9a1d-f95c000de11a-kube-api-access-r92x7\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.966138 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:32 crc kubenswrapper[4871]: I1008 00:43:32.966151 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426135f0-8fe9-40e8-9a1d-f95c000de11a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.636673 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"426135f0-8fe9-40e8-9a1d-f95c000de11a","Type":"ContainerDied","Data":"9da9e043413052afc73e2a81275d4e002e5fd4e60a5dd9833aa841028af4e81f"} Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.636721 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.637687 4871 scope.go:117] "RemoveContainer" containerID="463df52f01d8665c2d960be5d6a0bafa81b877c3ce9fde41cf2bbd3ab37dd67b" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.657739 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.673718 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.686168 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 00:43:33 crc kubenswrapper[4871]: E1008 00:43:33.686737 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426135f0-8fe9-40e8-9a1d-f95c000de11a" containerName="nova-cell0-conductor-conductor" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.686766 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="426135f0-8fe9-40e8-9a1d-f95c000de11a" containerName="nova-cell0-conductor-conductor" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.687074 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="426135f0-8fe9-40e8-9a1d-f95c000de11a" containerName="nova-cell0-conductor-conductor" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.688044 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.689885 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.700271 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.785655 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e5e4de2-9acb-4879-bba9-7466a6b77ba2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0e5e4de2-9acb-4879-bba9-7466a6b77ba2\") " pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.787571 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e5e4de2-9acb-4879-bba9-7466a6b77ba2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0e5e4de2-9acb-4879-bba9-7466a6b77ba2\") " pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.788158 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkxzb\" (UniqueName: \"kubernetes.io/projected/0e5e4de2-9acb-4879-bba9-7466a6b77ba2-kube-api-access-nkxzb\") pod \"nova-cell0-conductor-0\" (UID: \"0e5e4de2-9acb-4879-bba9-7466a6b77ba2\") " pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.890524 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e5e4de2-9acb-4879-bba9-7466a6b77ba2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0e5e4de2-9acb-4879-bba9-7466a6b77ba2\") " pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.890616 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkxzb\" (UniqueName: \"kubernetes.io/projected/0e5e4de2-9acb-4879-bba9-7466a6b77ba2-kube-api-access-nkxzb\") pod \"nova-cell0-conductor-0\" (UID: \"0e5e4de2-9acb-4879-bba9-7466a6b77ba2\") " pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:33 crc kubenswrapper[4871]: I1008 00:43:33.890688 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e5e4de2-9acb-4879-bba9-7466a6b77ba2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0e5e4de2-9acb-4879-bba9-7466a6b77ba2\") " pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:34 crc kubenswrapper[4871]: I1008 00:43:34.278072 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e5e4de2-9acb-4879-bba9-7466a6b77ba2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0e5e4de2-9acb-4879-bba9-7466a6b77ba2\") " pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:34 crc kubenswrapper[4871]: I1008 00:43:34.281085 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkxzb\" (UniqueName: \"kubernetes.io/projected/0e5e4de2-9acb-4879-bba9-7466a6b77ba2-kube-api-access-nkxzb\") pod \"nova-cell0-conductor-0\" (UID: \"0e5e4de2-9acb-4879-bba9-7466a6b77ba2\") " pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:34 crc kubenswrapper[4871]: I1008 00:43:34.291783 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e5e4de2-9acb-4879-bba9-7466a6b77ba2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0e5e4de2-9acb-4879-bba9-7466a6b77ba2\") " pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:34 crc kubenswrapper[4871]: I1008 00:43:34.314668 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:34 crc kubenswrapper[4871]: I1008 00:43:34.856061 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.004499 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="426135f0-8fe9-40e8-9a1d-f95c000de11a" path="/var/lib/kubelet/pods/426135f0-8fe9-40e8-9a1d-f95c000de11a/volumes" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.538219 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.661237 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0e5e4de2-9acb-4879-bba9-7466a6b77ba2","Type":"ContainerStarted","Data":"82369a5d204415e9c1d753d90268847262ec91bacec227abb1fc44d2c7c72d40"} Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.661277 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0e5e4de2-9acb-4879-bba9-7466a6b77ba2","Type":"ContainerStarted","Data":"d29bf4fb28d3369302a92df2c5c16745e473f20a266afd8dc7a21ac7d7010104"} Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.661300 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.662729 4871 generic.go:334] "Generic (PLEG): container finished" podID="a360e652-467f-4d96-9232-d944191ee751" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" exitCode=0 Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.662764 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a360e652-467f-4d96-9232-d944191ee751","Type":"ContainerDied","Data":"dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34"} Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.662808 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a360e652-467f-4d96-9232-d944191ee751","Type":"ContainerDied","Data":"47e2e8e455a3610896404f480ee6449179b946796104b0129f381165c5dc47e5"} Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.662829 4871 scope.go:117] "RemoveContainer" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.662961 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.663407 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-combined-ca-bundle\") pod \"a360e652-467f-4d96-9232-d944191ee751\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.663628 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqmqn\" (UniqueName: \"kubernetes.io/projected/a360e652-467f-4d96-9232-d944191ee751-kube-api-access-fqmqn\") pod \"a360e652-467f-4d96-9232-d944191ee751\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.663696 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-config-data\") pod \"a360e652-467f-4d96-9232-d944191ee751\" (UID: \"a360e652-467f-4d96-9232-d944191ee751\") " Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.671917 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a360e652-467f-4d96-9232-d944191ee751-kube-api-access-fqmqn" (OuterVolumeSpecName: "kube-api-access-fqmqn") pod "a360e652-467f-4d96-9232-d944191ee751" (UID: "a360e652-467f-4d96-9232-d944191ee751"). InnerVolumeSpecName "kube-api-access-fqmqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.681039 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.681019646 podStartE2EDuration="2.681019646s" podCreationTimestamp="2025-10-08 00:43:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:43:35.680726919 +0000 UTC m=+9289.483425032" watchObservedRunningTime="2025-10-08 00:43:35.681019646 +0000 UTC m=+9289.483717719" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.692634 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a360e652-467f-4d96-9232-d944191ee751" (UID: "a360e652-467f-4d96-9232-d944191ee751"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.721336 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-config-data" (OuterVolumeSpecName: "config-data") pod "a360e652-467f-4d96-9232-d944191ee751" (UID: "a360e652-467f-4d96-9232-d944191ee751"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.766378 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.766416 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqmqn\" (UniqueName: \"kubernetes.io/projected/a360e652-467f-4d96-9232-d944191ee751-kube-api-access-fqmqn\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.766429 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a360e652-467f-4d96-9232-d944191ee751-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.787158 4871 scope.go:117] "RemoveContainer" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" Oct 08 00:43:35 crc kubenswrapper[4871]: E1008 00:43:35.787630 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34\": container with ID starting with dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34 not found: ID does not exist" containerID="dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.787683 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34"} err="failed to get container status \"dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34\": rpc error: code = NotFound desc = could not find container \"dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34\": container with ID starting with dcbe47422205aa0117a890a6ab866ca1f30cd4e6784187a5affcd850c886ec34 not found: ID does not exist" Oct 08 00:43:35 crc kubenswrapper[4871]: I1008 00:43:35.999437 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.015606 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.028356 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 00:43:36 crc kubenswrapper[4871]: E1008 00:43:36.028947 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a360e652-467f-4d96-9232-d944191ee751" containerName="nova-scheduler-scheduler" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.028968 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="a360e652-467f-4d96-9232-d944191ee751" containerName="nova-scheduler-scheduler" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.029237 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="a360e652-467f-4d96-9232-d944191ee751" containerName="nova-scheduler-scheduler" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.030145 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.033099 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.039782 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.128987 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.129022 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.173932 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23fd31e-e2d3-4dfe-be60-e7497b9b6748-config-data\") pod \"nova-scheduler-0\" (UID: \"e23fd31e-e2d3-4dfe-be60-e7497b9b6748\") " pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.174551 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23fd31e-e2d3-4dfe-be60-e7497b9b6748-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e23fd31e-e2d3-4dfe-be60-e7497b9b6748\") " pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.174668 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvf4v\" (UniqueName: \"kubernetes.io/projected/e23fd31e-e2d3-4dfe-be60-e7497b9b6748-kube-api-access-qvf4v\") pod \"nova-scheduler-0\" (UID: \"e23fd31e-e2d3-4dfe-be60-e7497b9b6748\") " pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.276284 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23fd31e-e2d3-4dfe-be60-e7497b9b6748-config-data\") pod \"nova-scheduler-0\" (UID: \"e23fd31e-e2d3-4dfe-be60-e7497b9b6748\") " pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.276417 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23fd31e-e2d3-4dfe-be60-e7497b9b6748-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e23fd31e-e2d3-4dfe-be60-e7497b9b6748\") " pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.276445 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvf4v\" (UniqueName: \"kubernetes.io/projected/e23fd31e-e2d3-4dfe-be60-e7497b9b6748-kube-api-access-qvf4v\") pod \"nova-scheduler-0\" (UID: \"e23fd31e-e2d3-4dfe-be60-e7497b9b6748\") " pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.594167 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23fd31e-e2d3-4dfe-be60-e7497b9b6748-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e23fd31e-e2d3-4dfe-be60-e7497b9b6748\") " pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.594272 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23fd31e-e2d3-4dfe-be60-e7497b9b6748-config-data\") pod \"nova-scheduler-0\" (UID: \"e23fd31e-e2d3-4dfe-be60-e7497b9b6748\") " pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.594357 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvf4v\" (UniqueName: \"kubernetes.io/projected/e23fd31e-e2d3-4dfe-be60-e7497b9b6748-kube-api-access-qvf4v\") pod \"nova-scheduler-0\" (UID: \"e23fd31e-e2d3-4dfe-be60-e7497b9b6748\") " pod="openstack/nova-scheduler-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.646102 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 08 00:43:36 crc kubenswrapper[4871]: I1008 00:43:36.655468 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 00:43:37 crc kubenswrapper[4871]: I1008 00:43:37.002431 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a360e652-467f-4d96-9232-d944191ee751" path="/var/lib/kubelet/pods/a360e652-467f-4d96-9232-d944191ee751/volumes" Oct 08 00:43:37 crc kubenswrapper[4871]: I1008 00:43:37.164444 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 00:43:37 crc kubenswrapper[4871]: I1008 00:43:37.690651 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e23fd31e-e2d3-4dfe-be60-e7497b9b6748","Type":"ContainerStarted","Data":"f6a904fe885b8184b65afd93eb1911599485a2a13201968bd0b17ecb7290521c"} Oct 08 00:43:37 crc kubenswrapper[4871]: I1008 00:43:37.690938 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e23fd31e-e2d3-4dfe-be60-e7497b9b6748","Type":"ContainerStarted","Data":"9c3634be5d87addbb47e547156e22c6c6b1ad16d79810f4bf12269b15fa2b4b2"} Oct 08 00:43:37 crc kubenswrapper[4871]: I1008 00:43:37.718491 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.718469762 podStartE2EDuration="2.718469762s" podCreationTimestamp="2025-10-08 00:43:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:43:37.709179379 +0000 UTC m=+9291.511877452" watchObservedRunningTime="2025-10-08 00:43:37.718469762 +0000 UTC m=+9291.521167855" Oct 08 00:43:41 crc kubenswrapper[4871]: I1008 00:43:41.034989 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 00:43:41 crc kubenswrapper[4871]: I1008 00:43:41.035621 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 00:43:41 crc kubenswrapper[4871]: I1008 00:43:41.129030 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 00:43:41 crc kubenswrapper[4871]: I1008 00:43:41.129093 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 00:43:41 crc kubenswrapper[4871]: I1008 00:43:41.656354 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 00:43:42 crc kubenswrapper[4871]: I1008 00:43:42.117107 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d9f07d2a-d7a7-46b1-9e36-faff79d08add" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 00:43:42 crc kubenswrapper[4871]: I1008 00:43:42.117144 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d9f07d2a-d7a7-46b1-9e36-faff79d08add" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 00:43:42 crc kubenswrapper[4871]: I1008 00:43:42.211067 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4e277fae-f0f9-469b-87da-b7c88f40871f" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.195:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 00:43:42 crc kubenswrapper[4871]: I1008 00:43:42.211535 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4e277fae-f0f9-469b-87da-b7c88f40871f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.195:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 00:43:42 crc kubenswrapper[4871]: I1008 00:43:42.985846 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:43:42 crc kubenswrapper[4871]: E1008 00:43:42.986349 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:43:44 crc kubenswrapper[4871]: I1008 00:43:44.370228 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 08 00:43:46 crc kubenswrapper[4871]: I1008 00:43:46.656503 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 00:43:46 crc kubenswrapper[4871]: I1008 00:43:46.711017 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 00:43:46 crc kubenswrapper[4871]: I1008 00:43:46.843698 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.038333 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.039357 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.039472 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.041960 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.131318 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.132074 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.133189 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.858536 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.861128 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 00:43:51 crc kubenswrapper[4871]: I1008 00:43:51.861718 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 00:43:54 crc kubenswrapper[4871]: I1008 00:43:54.983451 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:43:54 crc kubenswrapper[4871]: E1008 00:43:54.985497 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:44:07 crc kubenswrapper[4871]: I1008 00:44:07.982018 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:44:07 crc kubenswrapper[4871]: E1008 00:44:07.982906 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:44:22 crc kubenswrapper[4871]: I1008 00:44:22.984151 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:44:22 crc kubenswrapper[4871]: E1008 00:44:22.985029 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:44:37 crc kubenswrapper[4871]: I1008 00:44:37.988678 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:44:37 crc kubenswrapper[4871]: E1008 00:44:37.991061 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:44:50 crc kubenswrapper[4871]: I1008 00:44:50.983976 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:44:50 crc kubenswrapper[4871]: E1008 00:44:50.984810 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.168830 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq"] Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.172776 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.176373 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.176613 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.197321 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq"] Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.251754 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d20b898c-b403-4b03-8743-c9b21afd1f6e-secret-volume\") pod \"collect-profiles-29331405-lnxtq\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.252660 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lthx\" (UniqueName: \"kubernetes.io/projected/d20b898c-b403-4b03-8743-c9b21afd1f6e-kube-api-access-6lthx\") pod \"collect-profiles-29331405-lnxtq\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.253000 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d20b898c-b403-4b03-8743-c9b21afd1f6e-config-volume\") pod \"collect-profiles-29331405-lnxtq\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.355918 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d20b898c-b403-4b03-8743-c9b21afd1f6e-secret-volume\") pod \"collect-profiles-29331405-lnxtq\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.356080 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lthx\" (UniqueName: \"kubernetes.io/projected/d20b898c-b403-4b03-8743-c9b21afd1f6e-kube-api-access-6lthx\") pod \"collect-profiles-29331405-lnxtq\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.356177 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d20b898c-b403-4b03-8743-c9b21afd1f6e-config-volume\") pod \"collect-profiles-29331405-lnxtq\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.357168 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d20b898c-b403-4b03-8743-c9b21afd1f6e-config-volume\") pod \"collect-profiles-29331405-lnxtq\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.363599 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d20b898c-b403-4b03-8743-c9b21afd1f6e-secret-volume\") pod \"collect-profiles-29331405-lnxtq\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.377440 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lthx\" (UniqueName: \"kubernetes.io/projected/d20b898c-b403-4b03-8743-c9b21afd1f6e-kube-api-access-6lthx\") pod \"collect-profiles-29331405-lnxtq\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.521067 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:00 crc kubenswrapper[4871]: I1008 00:45:00.986764 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq"] Oct 08 00:45:01 crc kubenswrapper[4871]: I1008 00:45:01.751599 4871 generic.go:334] "Generic (PLEG): container finished" podID="d20b898c-b403-4b03-8743-c9b21afd1f6e" containerID="7c001ac8e81be5239f5977231b40c64d3738d7e27b5826e39516983612414f44" exitCode=0 Oct 08 00:45:01 crc kubenswrapper[4871]: I1008 00:45:01.751641 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" event={"ID":"d20b898c-b403-4b03-8743-c9b21afd1f6e","Type":"ContainerDied","Data":"7c001ac8e81be5239f5977231b40c64d3738d7e27b5826e39516983612414f44"} Oct 08 00:45:01 crc kubenswrapper[4871]: I1008 00:45:01.752027 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" event={"ID":"d20b898c-b403-4b03-8743-c9b21afd1f6e","Type":"ContainerStarted","Data":"e675484993fe4594ec8b4d1cef889081369f9075beb54433590fe46a811dd4a1"} Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.526424 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.628016 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d20b898c-b403-4b03-8743-c9b21afd1f6e-config-volume\") pod \"d20b898c-b403-4b03-8743-c9b21afd1f6e\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.628241 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d20b898c-b403-4b03-8743-c9b21afd1f6e-secret-volume\") pod \"d20b898c-b403-4b03-8743-c9b21afd1f6e\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.628337 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lthx\" (UniqueName: \"kubernetes.io/projected/d20b898c-b403-4b03-8743-c9b21afd1f6e-kube-api-access-6lthx\") pod \"d20b898c-b403-4b03-8743-c9b21afd1f6e\" (UID: \"d20b898c-b403-4b03-8743-c9b21afd1f6e\") " Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.629037 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d20b898c-b403-4b03-8743-c9b21afd1f6e-config-volume" (OuterVolumeSpecName: "config-volume") pod "d20b898c-b403-4b03-8743-c9b21afd1f6e" (UID: "d20b898c-b403-4b03-8743-c9b21afd1f6e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.629538 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d20b898c-b403-4b03-8743-c9b21afd1f6e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.633523 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20b898c-b403-4b03-8743-c9b21afd1f6e-kube-api-access-6lthx" (OuterVolumeSpecName: "kube-api-access-6lthx") pod "d20b898c-b403-4b03-8743-c9b21afd1f6e" (UID: "d20b898c-b403-4b03-8743-c9b21afd1f6e"). InnerVolumeSpecName "kube-api-access-6lthx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.633568 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20b898c-b403-4b03-8743-c9b21afd1f6e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d20b898c-b403-4b03-8743-c9b21afd1f6e" (UID: "d20b898c-b403-4b03-8743-c9b21afd1f6e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.731913 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d20b898c-b403-4b03-8743-c9b21afd1f6e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.731945 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lthx\" (UniqueName: \"kubernetes.io/projected/d20b898c-b403-4b03-8743-c9b21afd1f6e-kube-api-access-6lthx\") on node \"crc\" DevicePath \"\"" Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.775081 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" event={"ID":"d20b898c-b403-4b03-8743-c9b21afd1f6e","Type":"ContainerDied","Data":"e675484993fe4594ec8b4d1cef889081369f9075beb54433590fe46a811dd4a1"} Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.775145 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e675484993fe4594ec8b4d1cef889081369f9075beb54433590fe46a811dd4a1" Oct 08 00:45:03 crc kubenswrapper[4871]: I1008 00:45:03.775151 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331405-lnxtq" Oct 08 00:45:04 crc kubenswrapper[4871]: I1008 00:45:04.630743 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7"] Oct 08 00:45:04 crc kubenswrapper[4871]: I1008 00:45:04.642872 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331360-847m7"] Oct 08 00:45:05 crc kubenswrapper[4871]: I1008 00:45:05.003921 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779" path="/var/lib/kubelet/pods/bcca9d1a-a6c4-47ef-bd7a-5c6d1b149779/volumes" Oct 08 00:45:05 crc kubenswrapper[4871]: I1008 00:45:05.982286 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:45:05 crc kubenswrapper[4871]: E1008 00:45:05.982626 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:45:17 crc kubenswrapper[4871]: I1008 00:45:17.007623 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:45:17 crc kubenswrapper[4871]: E1008 00:45:17.008461 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:45:28 crc kubenswrapper[4871]: I1008 00:45:28.982973 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:45:28 crc kubenswrapper[4871]: E1008 00:45:28.983860 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:45:41 crc kubenswrapper[4871]: I1008 00:45:41.982743 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:45:41 crc kubenswrapper[4871]: E1008 00:45:41.983511 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:45:54 crc kubenswrapper[4871]: I1008 00:45:54.983465 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:45:54 crc kubenswrapper[4871]: E1008 00:45:54.984656 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:46:04 crc kubenswrapper[4871]: I1008 00:46:04.065149 4871 scope.go:117] "RemoveContainer" containerID="65c58dd948bc3c17b2b9f07eb8e6e9f603c4b22926aa2f24fced696ab5d766d8" Oct 08 00:46:07 crc kubenswrapper[4871]: I1008 00:46:07.983195 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:46:08 crc kubenswrapper[4871]: I1008 00:46:08.652401 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"46ce7aa44079bc0a740013a3ef6a3d2367f1a30958b4917112276718bae9796b"} Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.129423 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gmlzb"] Oct 08 00:47:36 crc kubenswrapper[4871]: E1008 00:47:36.130537 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20b898c-b403-4b03-8743-c9b21afd1f6e" containerName="collect-profiles" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.130554 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20b898c-b403-4b03-8743-c9b21afd1f6e" containerName="collect-profiles" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.130813 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20b898c-b403-4b03-8743-c9b21afd1f6e" containerName="collect-profiles" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.132687 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.141539 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gmlzb"] Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.234309 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-catalog-content\") pod \"certified-operators-gmlzb\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.234640 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-utilities\") pod \"certified-operators-gmlzb\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.234692 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thdlt\" (UniqueName: \"kubernetes.io/projected/63ec9559-589a-4f9b-940a-b76608c844ee-kube-api-access-thdlt\") pod \"certified-operators-gmlzb\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.336882 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-utilities\") pod \"certified-operators-gmlzb\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.337262 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thdlt\" (UniqueName: \"kubernetes.io/projected/63ec9559-589a-4f9b-940a-b76608c844ee-kube-api-access-thdlt\") pod \"certified-operators-gmlzb\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.337353 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-utilities\") pod \"certified-operators-gmlzb\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.337403 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-catalog-content\") pod \"certified-operators-gmlzb\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.337660 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-catalog-content\") pod \"certified-operators-gmlzb\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.365551 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thdlt\" (UniqueName: \"kubernetes.io/projected/63ec9559-589a-4f9b-940a-b76608c844ee-kube-api-access-thdlt\") pod \"certified-operators-gmlzb\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:36 crc kubenswrapper[4871]: I1008 00:47:36.465715 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:37 crc kubenswrapper[4871]: I1008 00:47:37.060220 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gmlzb"] Oct 08 00:47:37 crc kubenswrapper[4871]: I1008 00:47:37.774260 4871 generic.go:334] "Generic (PLEG): container finished" podID="63ec9559-589a-4f9b-940a-b76608c844ee" containerID="fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab" exitCode=0 Oct 08 00:47:37 crc kubenswrapper[4871]: I1008 00:47:37.774506 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmlzb" event={"ID":"63ec9559-589a-4f9b-940a-b76608c844ee","Type":"ContainerDied","Data":"fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab"} Oct 08 00:47:37 crc kubenswrapper[4871]: I1008 00:47:37.775544 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmlzb" event={"ID":"63ec9559-589a-4f9b-940a-b76608c844ee","Type":"ContainerStarted","Data":"8a5684cc5e8700c7b9059d5fdd1cc9f02304a10641b1ed76f0d9d7eff88bd225"} Oct 08 00:47:38 crc kubenswrapper[4871]: I1008 00:47:38.788054 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmlzb" event={"ID":"63ec9559-589a-4f9b-940a-b76608c844ee","Type":"ContainerStarted","Data":"41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a"} Oct 08 00:47:39 crc kubenswrapper[4871]: I1008 00:47:39.804002 4871 generic.go:334] "Generic (PLEG): container finished" podID="63ec9559-589a-4f9b-940a-b76608c844ee" containerID="41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a" exitCode=0 Oct 08 00:47:39 crc kubenswrapper[4871]: I1008 00:47:39.804125 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmlzb" event={"ID":"63ec9559-589a-4f9b-940a-b76608c844ee","Type":"ContainerDied","Data":"41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a"} Oct 08 00:47:40 crc kubenswrapper[4871]: I1008 00:47:40.817971 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmlzb" event={"ID":"63ec9559-589a-4f9b-940a-b76608c844ee","Type":"ContainerStarted","Data":"a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078"} Oct 08 00:47:40 crc kubenswrapper[4871]: I1008 00:47:40.852224 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gmlzb" podStartSLOduration=2.39886507 podStartE2EDuration="4.85220618s" podCreationTimestamp="2025-10-08 00:47:36 +0000 UTC" firstStartedPulling="2025-10-08 00:47:37.777271371 +0000 UTC m=+9531.579969434" lastFinishedPulling="2025-10-08 00:47:40.230612481 +0000 UTC m=+9534.033310544" observedRunningTime="2025-10-08 00:47:40.851994455 +0000 UTC m=+9534.654692528" watchObservedRunningTime="2025-10-08 00:47:40.85220618 +0000 UTC m=+9534.654904243" Oct 08 00:47:46 crc kubenswrapper[4871]: I1008 00:47:46.465911 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:46 crc kubenswrapper[4871]: I1008 00:47:46.466479 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:46 crc kubenswrapper[4871]: I1008 00:47:46.511900 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:47 crc kubenswrapper[4871]: I1008 00:47:47.002087 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:47 crc kubenswrapper[4871]: I1008 00:47:47.053567 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gmlzb"] Oct 08 00:47:48 crc kubenswrapper[4871]: I1008 00:47:48.939484 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gmlzb" podUID="63ec9559-589a-4f9b-940a-b76608c844ee" containerName="registry-server" containerID="cri-o://a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078" gracePeriod=2 Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.423978 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.571364 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thdlt\" (UniqueName: \"kubernetes.io/projected/63ec9559-589a-4f9b-940a-b76608c844ee-kube-api-access-thdlt\") pod \"63ec9559-589a-4f9b-940a-b76608c844ee\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.571442 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-catalog-content\") pod \"63ec9559-589a-4f9b-940a-b76608c844ee\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.571706 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-utilities\") pod \"63ec9559-589a-4f9b-940a-b76608c844ee\" (UID: \"63ec9559-589a-4f9b-940a-b76608c844ee\") " Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.573220 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-utilities" (OuterVolumeSpecName: "utilities") pod "63ec9559-589a-4f9b-940a-b76608c844ee" (UID: "63ec9559-589a-4f9b-940a-b76608c844ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.580994 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ec9559-589a-4f9b-940a-b76608c844ee-kube-api-access-thdlt" (OuterVolumeSpecName: "kube-api-access-thdlt") pod "63ec9559-589a-4f9b-940a-b76608c844ee" (UID: "63ec9559-589a-4f9b-940a-b76608c844ee"). InnerVolumeSpecName "kube-api-access-thdlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.637673 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63ec9559-589a-4f9b-940a-b76608c844ee" (UID: "63ec9559-589a-4f9b-940a-b76608c844ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.673981 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.674038 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thdlt\" (UniqueName: \"kubernetes.io/projected/63ec9559-589a-4f9b-940a-b76608c844ee-kube-api-access-thdlt\") on node \"crc\" DevicePath \"\"" Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.674065 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ec9559-589a-4f9b-940a-b76608c844ee-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.954718 4871 generic.go:334] "Generic (PLEG): container finished" podID="63ec9559-589a-4f9b-940a-b76608c844ee" containerID="a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078" exitCode=0 Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.954830 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmlzb" event={"ID":"63ec9559-589a-4f9b-940a-b76608c844ee","Type":"ContainerDied","Data":"a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078"} Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.954873 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmlzb" event={"ID":"63ec9559-589a-4f9b-940a-b76608c844ee","Type":"ContainerDied","Data":"8a5684cc5e8700c7b9059d5fdd1cc9f02304a10641b1ed76f0d9d7eff88bd225"} Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.954903 4871 scope.go:117] "RemoveContainer" containerID="a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078" Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.955121 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmlzb" Oct 08 00:47:49 crc kubenswrapper[4871]: I1008 00:47:49.998123 4871 scope.go:117] "RemoveContainer" containerID="41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a" Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.012941 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gmlzb"] Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.026908 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gmlzb"] Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.032074 4871 scope.go:117] "RemoveContainer" containerID="fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab" Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.075321 4871 scope.go:117] "RemoveContainer" containerID="a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078" Oct 08 00:47:50 crc kubenswrapper[4871]: E1008 00:47:50.075868 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078\": container with ID starting with a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078 not found: ID does not exist" containerID="a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078" Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.075911 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078"} err="failed to get container status \"a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078\": rpc error: code = NotFound desc = could not find container \"a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078\": container with ID starting with a9b0974cd07a4564fef9f61ff253151f9645be5a9f80b7ccdb6eb0331480b078 not found: ID does not exist" Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.075938 4871 scope.go:117] "RemoveContainer" containerID="41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a" Oct 08 00:47:50 crc kubenswrapper[4871]: E1008 00:47:50.076190 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a\": container with ID starting with 41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a not found: ID does not exist" containerID="41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a" Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.076323 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a"} err="failed to get container status \"41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a\": rpc error: code = NotFound desc = could not find container \"41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a\": container with ID starting with 41aac08d080dd6f653b29c901e7e3798ed98feb8a75c3f99f94dde694ac85b4a not found: ID does not exist" Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.076345 4871 scope.go:117] "RemoveContainer" containerID="fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab" Oct 08 00:47:50 crc kubenswrapper[4871]: E1008 00:47:50.076657 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab\": container with ID starting with fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab not found: ID does not exist" containerID="fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab" Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.076687 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab"} err="failed to get container status \"fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab\": rpc error: code = NotFound desc = could not find container \"fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab\": container with ID starting with fecaed9fc3d4d1142801c05bee765cc43df497dd609bf04b79fe90e9008e4eab not found: ID does not exist" Oct 08 00:47:50 crc kubenswrapper[4871]: I1008 00:47:50.998316 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63ec9559-589a-4f9b-940a-b76608c844ee" path="/var/lib/kubelet/pods/63ec9559-589a-4f9b-940a-b76608c844ee/volumes" Oct 08 00:48:35 crc kubenswrapper[4871]: I1008 00:48:35.513845 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:48:35 crc kubenswrapper[4871]: I1008 00:48:35.514319 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:49:05 crc kubenswrapper[4871]: I1008 00:49:05.512960 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:49:05 crc kubenswrapper[4871]: I1008 00:49:05.513468 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:49:35 crc kubenswrapper[4871]: I1008 00:49:35.512058 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:49:35 crc kubenswrapper[4871]: I1008 00:49:35.512654 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:49:35 crc kubenswrapper[4871]: I1008 00:49:35.512701 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:49:35 crc kubenswrapper[4871]: I1008 00:49:35.513606 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46ce7aa44079bc0a740013a3ef6a3d2367f1a30958b4917112276718bae9796b"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:49:35 crc kubenswrapper[4871]: I1008 00:49:35.513679 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://46ce7aa44079bc0a740013a3ef6a3d2367f1a30958b4917112276718bae9796b" gracePeriod=600 Oct 08 00:49:36 crc kubenswrapper[4871]: I1008 00:49:36.400524 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="46ce7aa44079bc0a740013a3ef6a3d2367f1a30958b4917112276718bae9796b" exitCode=0 Oct 08 00:49:36 crc kubenswrapper[4871]: I1008 00:49:36.400624 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"46ce7aa44079bc0a740013a3ef6a3d2367f1a30958b4917112276718bae9796b"} Oct 08 00:49:36 crc kubenswrapper[4871]: I1008 00:49:36.400877 4871 scope.go:117] "RemoveContainer" containerID="99ac5dba5fa0bc704cf385feebcd19c645d4789c85e1f2b8c7506f13b211d877" Oct 08 00:49:37 crc kubenswrapper[4871]: I1008 00:49:37.412180 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807"} Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.347865 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cgxqx"] Oct 08 00:52:00 crc kubenswrapper[4871]: E1008 00:52:00.348901 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ec9559-589a-4f9b-940a-b76608c844ee" containerName="extract-utilities" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.348914 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ec9559-589a-4f9b-940a-b76608c844ee" containerName="extract-utilities" Oct 08 00:52:00 crc kubenswrapper[4871]: E1008 00:52:00.348941 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ec9559-589a-4f9b-940a-b76608c844ee" containerName="registry-server" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.348948 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ec9559-589a-4f9b-940a-b76608c844ee" containerName="registry-server" Oct 08 00:52:00 crc kubenswrapper[4871]: E1008 00:52:00.348959 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ec9559-589a-4f9b-940a-b76608c844ee" containerName="extract-content" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.348965 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ec9559-589a-4f9b-940a-b76608c844ee" containerName="extract-content" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.349171 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ec9559-589a-4f9b-940a-b76608c844ee" containerName="registry-server" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.350914 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.360286 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cgxqx"] Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.452508 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-catalog-content\") pod \"redhat-operators-cgxqx\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.452569 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cp4n\" (UniqueName: \"kubernetes.io/projected/7b35d72a-17d2-442e-a296-122cbe3bbc9b-kube-api-access-4cp4n\") pod \"redhat-operators-cgxqx\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.453170 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-utilities\") pod \"redhat-operators-cgxqx\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.555040 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-utilities\") pod \"redhat-operators-cgxqx\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.555181 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-catalog-content\") pod \"redhat-operators-cgxqx\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.555613 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-utilities\") pod \"redhat-operators-cgxqx\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.555641 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-catalog-content\") pod \"redhat-operators-cgxqx\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.555676 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cp4n\" (UniqueName: \"kubernetes.io/projected/7b35d72a-17d2-442e-a296-122cbe3bbc9b-kube-api-access-4cp4n\") pod \"redhat-operators-cgxqx\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.576839 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cp4n\" (UniqueName: \"kubernetes.io/projected/7b35d72a-17d2-442e-a296-122cbe3bbc9b-kube-api-access-4cp4n\") pod \"redhat-operators-cgxqx\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:00 crc kubenswrapper[4871]: I1008 00:52:00.682632 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:01 crc kubenswrapper[4871]: I1008 00:52:01.262718 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cgxqx"] Oct 08 00:52:02 crc kubenswrapper[4871]: I1008 00:52:02.281545 4871 generic.go:334] "Generic (PLEG): container finished" podID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerID="2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065" exitCode=0 Oct 08 00:52:02 crc kubenswrapper[4871]: I1008 00:52:02.281955 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgxqx" event={"ID":"7b35d72a-17d2-442e-a296-122cbe3bbc9b","Type":"ContainerDied","Data":"2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065"} Oct 08 00:52:02 crc kubenswrapper[4871]: I1008 00:52:02.281981 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgxqx" event={"ID":"7b35d72a-17d2-442e-a296-122cbe3bbc9b","Type":"ContainerStarted","Data":"52d77b87783ac577b2e2d9569657f74e4a16439dfa356d80ce1c81870c6a6dc7"} Oct 08 00:52:02 crc kubenswrapper[4871]: I1008 00:52:02.283565 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:52:05 crc kubenswrapper[4871]: I1008 00:52:05.318721 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgxqx" event={"ID":"7b35d72a-17d2-442e-a296-122cbe3bbc9b","Type":"ContainerStarted","Data":"c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369"} Oct 08 00:52:05 crc kubenswrapper[4871]: I1008 00:52:05.512393 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:52:05 crc kubenswrapper[4871]: I1008 00:52:05.512469 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:52:06 crc kubenswrapper[4871]: I1008 00:52:06.342024 4871 generic.go:334] "Generic (PLEG): container finished" podID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerID="c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369" exitCode=0 Oct 08 00:52:06 crc kubenswrapper[4871]: I1008 00:52:06.342110 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgxqx" event={"ID":"7b35d72a-17d2-442e-a296-122cbe3bbc9b","Type":"ContainerDied","Data":"c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369"} Oct 08 00:52:08 crc kubenswrapper[4871]: I1008 00:52:08.366837 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgxqx" event={"ID":"7b35d72a-17d2-442e-a296-122cbe3bbc9b","Type":"ContainerStarted","Data":"65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff"} Oct 08 00:52:08 crc kubenswrapper[4871]: I1008 00:52:08.398296 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cgxqx" podStartSLOduration=3.369303642 podStartE2EDuration="8.398279013s" podCreationTimestamp="2025-10-08 00:52:00 +0000 UTC" firstStartedPulling="2025-10-08 00:52:02.283313889 +0000 UTC m=+9796.086011962" lastFinishedPulling="2025-10-08 00:52:07.31228922 +0000 UTC m=+9801.114987333" observedRunningTime="2025-10-08 00:52:08.385847418 +0000 UTC m=+9802.188545491" watchObservedRunningTime="2025-10-08 00:52:08.398279013 +0000 UTC m=+9802.200977086" Oct 08 00:52:10 crc kubenswrapper[4871]: I1008 00:52:10.683106 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:10 crc kubenswrapper[4871]: I1008 00:52:10.683905 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:11 crc kubenswrapper[4871]: I1008 00:52:11.737380 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cgxqx" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerName="registry-server" probeResult="failure" output=< Oct 08 00:52:11 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 08 00:52:11 crc kubenswrapper[4871]: > Oct 08 00:52:20 crc kubenswrapper[4871]: I1008 00:52:20.760512 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:20 crc kubenswrapper[4871]: I1008 00:52:20.835561 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:21 crc kubenswrapper[4871]: I1008 00:52:21.010282 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cgxqx"] Oct 08 00:52:22 crc kubenswrapper[4871]: I1008 00:52:22.515951 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cgxqx" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerName="registry-server" containerID="cri-o://65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff" gracePeriod=2 Oct 08 00:52:22 crc kubenswrapper[4871]: E1008 00:52:22.898135 4871 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b35d72a_17d2_442e_a296_122cbe3bbc9b.slice/crio-conmon-65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b35d72a_17d2_442e_a296_122cbe3bbc9b.slice/crio-65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff.scope\": RecentStats: unable to find data in memory cache]" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.158506 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.199502 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-catalog-content\") pod \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.199715 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-utilities\") pod \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.199858 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cp4n\" (UniqueName: \"kubernetes.io/projected/7b35d72a-17d2-442e-a296-122cbe3bbc9b-kube-api-access-4cp4n\") pod \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\" (UID: \"7b35d72a-17d2-442e-a296-122cbe3bbc9b\") " Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.200686 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-utilities" (OuterVolumeSpecName: "utilities") pod "7b35d72a-17d2-442e-a296-122cbe3bbc9b" (UID: "7b35d72a-17d2-442e-a296-122cbe3bbc9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.209010 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b35d72a-17d2-442e-a296-122cbe3bbc9b-kube-api-access-4cp4n" (OuterVolumeSpecName: "kube-api-access-4cp4n") pod "7b35d72a-17d2-442e-a296-122cbe3bbc9b" (UID: "7b35d72a-17d2-442e-a296-122cbe3bbc9b"). InnerVolumeSpecName "kube-api-access-4cp4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.289486 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b35d72a-17d2-442e-a296-122cbe3bbc9b" (UID: "7b35d72a-17d2-442e-a296-122cbe3bbc9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.302261 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.302302 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b35d72a-17d2-442e-a296-122cbe3bbc9b-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.302319 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cp4n\" (UniqueName: \"kubernetes.io/projected/7b35d72a-17d2-442e-a296-122cbe3bbc9b-kube-api-access-4cp4n\") on node \"crc\" DevicePath \"\"" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.533779 4871 generic.go:334] "Generic (PLEG): container finished" podID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerID="65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff" exitCode=0 Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.533846 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgxqx" event={"ID":"7b35d72a-17d2-442e-a296-122cbe3bbc9b","Type":"ContainerDied","Data":"65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff"} Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.533878 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgxqx" event={"ID":"7b35d72a-17d2-442e-a296-122cbe3bbc9b","Type":"ContainerDied","Data":"52d77b87783ac577b2e2d9569657f74e4a16439dfa356d80ce1c81870c6a6dc7"} Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.533900 4871 scope.go:117] "RemoveContainer" containerID="65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.534053 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cgxqx" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.587033 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cgxqx"] Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.590977 4871 scope.go:117] "RemoveContainer" containerID="c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.599848 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cgxqx"] Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.620531 4871 scope.go:117] "RemoveContainer" containerID="2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.689432 4871 scope.go:117] "RemoveContainer" containerID="65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff" Oct 08 00:52:23 crc kubenswrapper[4871]: E1008 00:52:23.690382 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff\": container with ID starting with 65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff not found: ID does not exist" containerID="65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.690429 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff"} err="failed to get container status \"65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff\": rpc error: code = NotFound desc = could not find container \"65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff\": container with ID starting with 65d873ee77af833a83e93f35a3caabdebf63ac0593836354695e0af1d911f6ff not found: ID does not exist" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.690455 4871 scope.go:117] "RemoveContainer" containerID="c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369" Oct 08 00:52:23 crc kubenswrapper[4871]: E1008 00:52:23.691058 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369\": container with ID starting with c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369 not found: ID does not exist" containerID="c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.691109 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369"} err="failed to get container status \"c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369\": rpc error: code = NotFound desc = could not find container \"c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369\": container with ID starting with c2bb5fb2f633deb6d530ba60604fa8c62392119345fd4b25a81fe99d227a6369 not found: ID does not exist" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.691141 4871 scope.go:117] "RemoveContainer" containerID="2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065" Oct 08 00:52:23 crc kubenswrapper[4871]: E1008 00:52:23.691588 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065\": container with ID starting with 2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065 not found: ID does not exist" containerID="2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065" Oct 08 00:52:23 crc kubenswrapper[4871]: I1008 00:52:23.691610 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065"} err="failed to get container status \"2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065\": rpc error: code = NotFound desc = could not find container \"2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065\": container with ID starting with 2cce42ac8bb702b8b686d375ac65727fb4b1eabc940874e9510f15aec5252065 not found: ID does not exist" Oct 08 00:52:25 crc kubenswrapper[4871]: I1008 00:52:25.005718 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" path="/var/lib/kubelet/pods/7b35d72a-17d2-442e-a296-122cbe3bbc9b/volumes" Oct 08 00:52:35 crc kubenswrapper[4871]: I1008 00:52:35.512957 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:52:35 crc kubenswrapper[4871]: I1008 00:52:35.513786 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:53:05 crc kubenswrapper[4871]: I1008 00:53:05.073122 4871 generic.go:334] "Generic (PLEG): container finished" podID="b8b25c96-823b-433c-8512-cf364233ebe0" containerID="28948d188d5b1fdd0e9987a6cd018f729af05f8c6ecdfdbbaf9cb4a2cc25bbbe" exitCode=0 Oct 08 00:53:05 crc kubenswrapper[4871]: I1008 00:53:05.073267 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" event={"ID":"b8b25c96-823b-433c-8512-cf364233ebe0","Type":"ContainerDied","Data":"28948d188d5b1fdd0e9987a6cd018f729af05f8c6ecdfdbbaf9cb4a2cc25bbbe"} Oct 08 00:53:05 crc kubenswrapper[4871]: I1008 00:53:05.512285 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:53:05 crc kubenswrapper[4871]: I1008 00:53:05.512393 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:53:05 crc kubenswrapper[4871]: I1008 00:53:05.512476 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 00:53:05 crc kubenswrapper[4871]: I1008 00:53:05.514058 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:53:05 crc kubenswrapper[4871]: I1008 00:53:05.514207 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" gracePeriod=600 Oct 08 00:53:05 crc kubenswrapper[4871]: E1008 00:53:05.645524 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.092494 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" exitCode=0 Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.093381 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807"} Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.093450 4871 scope.go:117] "RemoveContainer" containerID="46ce7aa44079bc0a740013a3ef6a3d2367f1a30958b4917112276718bae9796b" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.094704 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:53:06 crc kubenswrapper[4871]: E1008 00:53:06.095342 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.544009 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.603270 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-0\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.603565 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68dzn\" (UniqueName: \"kubernetes.io/projected/b8b25c96-823b-433c-8512-cf364233ebe0-kube-api-access-68dzn\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.603657 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-1\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.603759 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-1\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.604024 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ssh-key\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.604265 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ceph\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.604862 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-0\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.604983 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-1\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.605109 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-combined-ca-bundle\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.605232 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-0\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.605454 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-inventory\") pod \"b8b25c96-823b-433c-8512-cf364233ebe0\" (UID: \"b8b25c96-823b-433c-8512-cf364233ebe0\") " Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.616670 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ceph" (OuterVolumeSpecName: "ceph") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.621902 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.626490 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8b25c96-823b-433c-8512-cf364233ebe0-kube-api-access-68dzn" (OuterVolumeSpecName: "kube-api-access-68dzn") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "kube-api-access-68dzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.643476 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.644271 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.650873 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-inventory" (OuterVolumeSpecName: "inventory") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.653577 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.653651 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.662017 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.664967 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.668669 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "b8b25c96-823b-433c-8512-cf364233ebe0" (UID: "b8b25c96-823b-433c-8512-cf364233ebe0"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719199 4871 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719237 4871 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719255 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719274 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719292 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719308 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719328 4871 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719345 4871 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719363 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68dzn\" (UniqueName: \"kubernetes.io/projected/b8b25c96-823b-433c-8512-cf364233ebe0-kube-api-access-68dzn\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719380 4871 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:06 crc kubenswrapper[4871]: I1008 00:53:06.719396 4871 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8b25c96-823b-433c-8512-cf364233ebe0-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:07 crc kubenswrapper[4871]: I1008 00:53:07.112214 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" event={"ID":"b8b25c96-823b-433c-8512-cf364233ebe0","Type":"ContainerDied","Data":"9f5217d4231849819da4780647f55450fad380b75fb745744ac4d6fd621aa466"} Oct 08 00:53:07 crc kubenswrapper[4871]: I1008 00:53:07.113709 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f5217d4231849819da4780647f55450fad380b75fb745744ac4d6fd621aa466" Oct 08 00:53:07 crc kubenswrapper[4871]: I1008 00:53:07.112337 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.879896 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hfvq8"] Oct 08 00:53:16 crc kubenswrapper[4871]: E1008 00:53:16.882004 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerName="extract-utilities" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.882022 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerName="extract-utilities" Oct 08 00:53:16 crc kubenswrapper[4871]: E1008 00:53:16.882046 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerName="registry-server" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.882053 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerName="registry-server" Oct 08 00:53:16 crc kubenswrapper[4871]: E1008 00:53:16.882071 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerName="extract-content" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.882078 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerName="extract-content" Oct 08 00:53:16 crc kubenswrapper[4871]: E1008 00:53:16.882091 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8b25c96-823b-433c-8512-cf364233ebe0" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.882099 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8b25c96-823b-433c-8512-cf364233ebe0" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.882402 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8b25c96-823b-433c-8512-cf364233ebe0" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.882425 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b35d72a-17d2-442e-a296-122cbe3bbc9b" containerName="registry-server" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.884942 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.893250 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hfvq8"] Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.958186 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-utilities\") pod \"community-operators-hfvq8\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.958244 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-catalog-content\") pod \"community-operators-hfvq8\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.958277 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmv8s\" (UniqueName: \"kubernetes.io/projected/fc3765dc-3b79-40e4-9dd2-03461978552f-kube-api-access-gmv8s\") pod \"community-operators-hfvq8\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:16 crc kubenswrapper[4871]: I1008 00:53:16.990013 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:53:16 crc kubenswrapper[4871]: E1008 00:53:16.990428 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:53:17 crc kubenswrapper[4871]: I1008 00:53:17.060193 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-utilities\") pod \"community-operators-hfvq8\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:17 crc kubenswrapper[4871]: I1008 00:53:17.060251 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-catalog-content\") pod \"community-operators-hfvq8\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:17 crc kubenswrapper[4871]: I1008 00:53:17.060285 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmv8s\" (UniqueName: \"kubernetes.io/projected/fc3765dc-3b79-40e4-9dd2-03461978552f-kube-api-access-gmv8s\") pod \"community-operators-hfvq8\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:17 crc kubenswrapper[4871]: I1008 00:53:17.060782 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-catalog-content\") pod \"community-operators-hfvq8\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:17 crc kubenswrapper[4871]: I1008 00:53:17.061002 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-utilities\") pod \"community-operators-hfvq8\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:17 crc kubenswrapper[4871]: I1008 00:53:17.089109 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmv8s\" (UniqueName: \"kubernetes.io/projected/fc3765dc-3b79-40e4-9dd2-03461978552f-kube-api-access-gmv8s\") pod \"community-operators-hfvq8\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:17 crc kubenswrapper[4871]: I1008 00:53:17.228641 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:17 crc kubenswrapper[4871]: I1008 00:53:17.784274 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hfvq8"] Oct 08 00:53:18 crc kubenswrapper[4871]: I1008 00:53:18.244943 4871 generic.go:334] "Generic (PLEG): container finished" podID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerID="e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a" exitCode=0 Oct 08 00:53:18 crc kubenswrapper[4871]: I1008 00:53:18.245024 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfvq8" event={"ID":"fc3765dc-3b79-40e4-9dd2-03461978552f","Type":"ContainerDied","Data":"e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a"} Oct 08 00:53:18 crc kubenswrapper[4871]: I1008 00:53:18.245269 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfvq8" event={"ID":"fc3765dc-3b79-40e4-9dd2-03461978552f","Type":"ContainerStarted","Data":"ddfdf451a01a40aa442536e888367bf4db1fb42426b4778596d02666936716f3"} Oct 08 00:53:19 crc kubenswrapper[4871]: I1008 00:53:19.259588 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfvq8" event={"ID":"fc3765dc-3b79-40e4-9dd2-03461978552f","Type":"ContainerStarted","Data":"0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446"} Oct 08 00:53:20 crc kubenswrapper[4871]: I1008 00:53:20.269900 4871 generic.go:334] "Generic (PLEG): container finished" podID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerID="0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446" exitCode=0 Oct 08 00:53:20 crc kubenswrapper[4871]: I1008 00:53:20.269981 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfvq8" event={"ID":"fc3765dc-3b79-40e4-9dd2-03461978552f","Type":"ContainerDied","Data":"0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446"} Oct 08 00:53:21 crc kubenswrapper[4871]: I1008 00:53:21.308627 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfvq8" event={"ID":"fc3765dc-3b79-40e4-9dd2-03461978552f","Type":"ContainerStarted","Data":"358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e"} Oct 08 00:53:21 crc kubenswrapper[4871]: I1008 00:53:21.335240 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hfvq8" podStartSLOduration=2.816038914 podStartE2EDuration="5.335221667s" podCreationTimestamp="2025-10-08 00:53:16 +0000 UTC" firstStartedPulling="2025-10-08 00:53:18.249173607 +0000 UTC m=+9872.051871680" lastFinishedPulling="2025-10-08 00:53:20.76835631 +0000 UTC m=+9874.571054433" observedRunningTime="2025-10-08 00:53:21.327626618 +0000 UTC m=+9875.130324701" watchObservedRunningTime="2025-10-08 00:53:21.335221667 +0000 UTC m=+9875.137919740" Oct 08 00:53:27 crc kubenswrapper[4871]: I1008 00:53:27.229042 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:27 crc kubenswrapper[4871]: I1008 00:53:27.229915 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:27 crc kubenswrapper[4871]: I1008 00:53:27.754033 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:27 crc kubenswrapper[4871]: I1008 00:53:27.830227 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:27 crc kubenswrapper[4871]: I1008 00:53:27.997027 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hfvq8"] Oct 08 00:53:29 crc kubenswrapper[4871]: I1008 00:53:29.401259 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hfvq8" podUID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerName="registry-server" containerID="cri-o://358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e" gracePeriod=2 Oct 08 00:53:29 crc kubenswrapper[4871]: I1008 00:53:29.975114 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.132784 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-utilities\") pod \"fc3765dc-3b79-40e4-9dd2-03461978552f\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.132907 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmv8s\" (UniqueName: \"kubernetes.io/projected/fc3765dc-3b79-40e4-9dd2-03461978552f-kube-api-access-gmv8s\") pod \"fc3765dc-3b79-40e4-9dd2-03461978552f\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.133198 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-catalog-content\") pod \"fc3765dc-3b79-40e4-9dd2-03461978552f\" (UID: \"fc3765dc-3b79-40e4-9dd2-03461978552f\") " Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.134815 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-utilities" (OuterVolumeSpecName: "utilities") pod "fc3765dc-3b79-40e4-9dd2-03461978552f" (UID: "fc3765dc-3b79-40e4-9dd2-03461978552f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.142971 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc3765dc-3b79-40e4-9dd2-03461978552f-kube-api-access-gmv8s" (OuterVolumeSpecName: "kube-api-access-gmv8s") pod "fc3765dc-3b79-40e4-9dd2-03461978552f" (UID: "fc3765dc-3b79-40e4-9dd2-03461978552f"). InnerVolumeSpecName "kube-api-access-gmv8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.208606 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc3765dc-3b79-40e4-9dd2-03461978552f" (UID: "fc3765dc-3b79-40e4-9dd2-03461978552f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.236645 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.236691 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc3765dc-3b79-40e4-9dd2-03461978552f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.236705 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmv8s\" (UniqueName: \"kubernetes.io/projected/fc3765dc-3b79-40e4-9dd2-03461978552f-kube-api-access-gmv8s\") on node \"crc\" DevicePath \"\"" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.414698 4871 generic.go:334] "Generic (PLEG): container finished" podID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerID="358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e" exitCode=0 Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.415075 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfvq8" event={"ID":"fc3765dc-3b79-40e4-9dd2-03461978552f","Type":"ContainerDied","Data":"358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e"} Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.415143 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfvq8" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.415172 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfvq8" event={"ID":"fc3765dc-3b79-40e4-9dd2-03461978552f","Type":"ContainerDied","Data":"ddfdf451a01a40aa442536e888367bf4db1fb42426b4778596d02666936716f3"} Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.415260 4871 scope.go:117] "RemoveContainer" containerID="358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.435988 4871 scope.go:117] "RemoveContainer" containerID="0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.453056 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hfvq8"] Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.462760 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hfvq8"] Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.482625 4871 scope.go:117] "RemoveContainer" containerID="e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.514820 4871 scope.go:117] "RemoveContainer" containerID="358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e" Oct 08 00:53:30 crc kubenswrapper[4871]: E1008 00:53:30.515288 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e\": container with ID starting with 358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e not found: ID does not exist" containerID="358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.515327 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e"} err="failed to get container status \"358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e\": rpc error: code = NotFound desc = could not find container \"358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e\": container with ID starting with 358a32a4b9da04f75f1a5af24ad7b7ec34d5d37f0819e915a5d6ca9d729c9f1e not found: ID does not exist" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.515348 4871 scope.go:117] "RemoveContainer" containerID="0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446" Oct 08 00:53:30 crc kubenswrapper[4871]: E1008 00:53:30.515595 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446\": container with ID starting with 0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446 not found: ID does not exist" containerID="0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.515615 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446"} err="failed to get container status \"0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446\": rpc error: code = NotFound desc = could not find container \"0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446\": container with ID starting with 0486d9b747e2ddeb7b0aa29cabd790eda34f852181669984782d5f400b232446 not found: ID does not exist" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.515630 4871 scope.go:117] "RemoveContainer" containerID="e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a" Oct 08 00:53:30 crc kubenswrapper[4871]: E1008 00:53:30.515909 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a\": container with ID starting with e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a not found: ID does not exist" containerID="e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.515928 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a"} err="failed to get container status \"e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a\": rpc error: code = NotFound desc = could not find container \"e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a\": container with ID starting with e9da64e4cb8d907cfb0adacf5b906b7e285401eee37c628da010773b7b561f1a not found: ID does not exist" Oct 08 00:53:30 crc kubenswrapper[4871]: I1008 00:53:30.998535 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc3765dc-3b79-40e4-9dd2-03461978552f" path="/var/lib/kubelet/pods/fc3765dc-3b79-40e4-9dd2-03461978552f/volumes" Oct 08 00:53:31 crc kubenswrapper[4871]: I1008 00:53:31.983707 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:53:31 crc kubenswrapper[4871]: E1008 00:53:31.984614 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:53:42 crc kubenswrapper[4871]: E1008 00:53:42.385092 4871 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.68:55868->38.102.83.68:41109: write tcp 38.102.83.68:55868->38.102.83.68:41109: write: broken pipe Oct 08 00:53:46 crc kubenswrapper[4871]: I1008 00:53:46.993500 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:53:46 crc kubenswrapper[4871]: E1008 00:53:46.994712 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:53:57 crc kubenswrapper[4871]: I1008 00:53:57.983734 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:53:57 crc kubenswrapper[4871]: E1008 00:53:57.984937 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:54:09 crc kubenswrapper[4871]: I1008 00:54:09.982993 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:54:09 crc kubenswrapper[4871]: E1008 00:54:09.984590 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:54:17 crc kubenswrapper[4871]: E1008 00:54:17.977290 4871 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.68:36288->38.102.83.68:41109: write tcp 38.102.83.68:36288->38.102.83.68:41109: write: broken pipe Oct 08 00:54:22 crc kubenswrapper[4871]: I1008 00:54:22.982513 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:54:22 crc kubenswrapper[4871]: E1008 00:54:22.983480 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:54:35 crc kubenswrapper[4871]: I1008 00:54:35.984024 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:54:35 crc kubenswrapper[4871]: E1008 00:54:35.985235 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:54:49 crc kubenswrapper[4871]: I1008 00:54:49.982445 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:54:49 crc kubenswrapper[4871]: E1008 00:54:49.983285 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:55:02 crc kubenswrapper[4871]: I1008 00:55:02.983287 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:55:02 crc kubenswrapper[4871]: E1008 00:55:02.985285 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:55:09 crc kubenswrapper[4871]: I1008 00:55:09.605200 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 08 00:55:09 crc kubenswrapper[4871]: I1008 00:55:09.606075 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="2c49b518-7037-4062-99ad-072d766d43a3" containerName="adoption" containerID="cri-o://ba2c3e5ec3b1e578580bf7573002682138386e24086f6b08f080ce48de024efa" gracePeriod=30 Oct 08 00:55:16 crc kubenswrapper[4871]: I1008 00:55:16.991358 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:55:16 crc kubenswrapper[4871]: E1008 00:55:16.992345 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:55:29 crc kubenswrapper[4871]: I1008 00:55:29.983482 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:55:29 crc kubenswrapper[4871]: E1008 00:55:29.984400 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:55:39 crc kubenswrapper[4871]: I1008 00:55:39.992203 4871 generic.go:334] "Generic (PLEG): container finished" podID="2c49b518-7037-4062-99ad-072d766d43a3" containerID="ba2c3e5ec3b1e578580bf7573002682138386e24086f6b08f080ce48de024efa" exitCode=137 Oct 08 00:55:39 crc kubenswrapper[4871]: I1008 00:55:39.992817 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"2c49b518-7037-4062-99ad-072d766d43a3","Type":"ContainerDied","Data":"ba2c3e5ec3b1e578580bf7573002682138386e24086f6b08f080ce48de024efa"} Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.195099 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.344787 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r22j\" (UniqueName: \"kubernetes.io/projected/2c49b518-7037-4062-99ad-072d766d43a3-kube-api-access-2r22j\") pod \"2c49b518-7037-4062-99ad-072d766d43a3\" (UID: \"2c49b518-7037-4062-99ad-072d766d43a3\") " Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.345848 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\") pod \"2c49b518-7037-4062-99ad-072d766d43a3\" (UID: \"2c49b518-7037-4062-99ad-072d766d43a3\") " Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.351430 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c49b518-7037-4062-99ad-072d766d43a3-kube-api-access-2r22j" (OuterVolumeSpecName: "kube-api-access-2r22j") pod "2c49b518-7037-4062-99ad-072d766d43a3" (UID: "2c49b518-7037-4062-99ad-072d766d43a3"). InnerVolumeSpecName "kube-api-access-2r22j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.384369 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0" (OuterVolumeSpecName: "mariadb-data") pod "2c49b518-7037-4062-99ad-072d766d43a3" (UID: "2c49b518-7037-4062-99ad-072d766d43a3"). InnerVolumeSpecName "pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.448490 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r22j\" (UniqueName: \"kubernetes.io/projected/2c49b518-7037-4062-99ad-072d766d43a3-kube-api-access-2r22j\") on node \"crc\" DevicePath \"\"" Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.448577 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\") on node \"crc\" " Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.485376 4871 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.485630 4871 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0") on node "crc" Oct 08 00:55:40 crc kubenswrapper[4871]: I1008 00:55:40.550463 4871 reconciler_common.go:293] "Volume detached for volume \"pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d5ad520-2f5e-4315-b0dd-b2eeb0a5f6f0\") on node \"crc\" DevicePath \"\"" Oct 08 00:55:41 crc kubenswrapper[4871]: I1008 00:55:41.006282 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"2c49b518-7037-4062-99ad-072d766d43a3","Type":"ContainerDied","Data":"b1c6e628fd76c5e078863763c51ee084754bb5149b6e6a64e6ae2987a7d1fc8f"} Oct 08 00:55:41 crc kubenswrapper[4871]: I1008 00:55:41.006351 4871 scope.go:117] "RemoveContainer" containerID="ba2c3e5ec3b1e578580bf7573002682138386e24086f6b08f080ce48de024efa" Oct 08 00:55:41 crc kubenswrapper[4871]: I1008 00:55:41.006396 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 08 00:55:41 crc kubenswrapper[4871]: I1008 00:55:41.038733 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 08 00:55:41 crc kubenswrapper[4871]: I1008 00:55:41.050651 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Oct 08 00:55:41 crc kubenswrapper[4871]: I1008 00:55:41.652093 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 08 00:55:41 crc kubenswrapper[4871]: I1008 00:55:41.652310 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="7b6206f2-ab54-4ec5-97e6-ed2460f6ee30" containerName="adoption" containerID="cri-o://76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455" gracePeriod=30 Oct 08 00:55:41 crc kubenswrapper[4871]: I1008 00:55:41.982668 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:55:41 crc kubenswrapper[4871]: E1008 00:55:41.982990 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:55:43 crc kubenswrapper[4871]: I1008 00:55:42.999688 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c49b518-7037-4062-99ad-072d766d43a3" path="/var/lib/kubelet/pods/2c49b518-7037-4062-99ad-072d766d43a3/volumes" Oct 08 00:55:56 crc kubenswrapper[4871]: I1008 00:55:56.992474 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:55:56 crc kubenswrapper[4871]: E1008 00:55:56.993324 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:56:08 crc kubenswrapper[4871]: I1008 00:56:08.982078 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:56:08 crc kubenswrapper[4871]: E1008 00:56:08.982837 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.279911 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.404075 4871 generic.go:334] "Generic (PLEG): container finished" podID="7b6206f2-ab54-4ec5-97e6-ed2460f6ee30" containerID="76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455" exitCode=137 Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.404128 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30","Type":"ContainerDied","Data":"76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455"} Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.404162 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30","Type":"ContainerDied","Data":"bfb19b21f38558b0a6e163e2a8be829a778c9f7dfc9ec71f2353acc99e1d8f42"} Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.404186 4871 scope.go:117] "RemoveContainer" containerID="76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.404211 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.437724 4871 scope.go:117] "RemoveContainer" containerID="76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455" Oct 08 00:56:12 crc kubenswrapper[4871]: E1008 00:56:12.439502 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455\": container with ID starting with 76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455 not found: ID does not exist" containerID="76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.439541 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455"} err="failed to get container status \"76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455\": rpc error: code = NotFound desc = could not find container \"76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455\": container with ID starting with 76d0fcc91c7a18d6f6e4b6442b47e546a58064623e295ff06414e09593f0b455 not found: ID does not exist" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.450480 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw4t7\" (UniqueName: \"kubernetes.io/projected/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-kube-api-access-qw4t7\") pod \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.451223 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\") pod \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.451280 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-ovn-data-cert\") pod \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\" (UID: \"7b6206f2-ab54-4ec5-97e6-ed2460f6ee30\") " Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.458371 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-kube-api-access-qw4t7" (OuterVolumeSpecName: "kube-api-access-qw4t7") pod "7b6206f2-ab54-4ec5-97e6-ed2460f6ee30" (UID: "7b6206f2-ab54-4ec5-97e6-ed2460f6ee30"). InnerVolumeSpecName "kube-api-access-qw4t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.458679 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "7b6206f2-ab54-4ec5-97e6-ed2460f6ee30" (UID: "7b6206f2-ab54-4ec5-97e6-ed2460f6ee30"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.476085 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f" (OuterVolumeSpecName: "ovn-data") pod "7b6206f2-ab54-4ec5-97e6-ed2460f6ee30" (UID: "7b6206f2-ab54-4ec5-97e6-ed2460f6ee30"). InnerVolumeSpecName "pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.555094 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw4t7\" (UniqueName: \"kubernetes.io/projected/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-kube-api-access-qw4t7\") on node \"crc\" DevicePath \"\"" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.555725 4871 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\") on node \"crc\" " Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.555752 4871 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.589064 4871 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.589341 4871 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f") on node "crc" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.657974 4871 reconciler_common.go:293] "Volume detached for volume \"pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9c04f3d-9b68-44a9-90b8-e7466677d77f\") on node \"crc\" DevicePath \"\"" Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.746023 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 08 00:56:12 crc kubenswrapper[4871]: I1008 00:56:12.756850 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Oct 08 00:56:13 crc kubenswrapper[4871]: I1008 00:56:12.999846 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b6206f2-ab54-4ec5-97e6-ed2460f6ee30" path="/var/lib/kubelet/pods/7b6206f2-ab54-4ec5-97e6-ed2460f6ee30/volumes" Oct 08 00:56:23 crc kubenswrapper[4871]: I1008 00:56:23.983466 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:56:23 crc kubenswrapper[4871]: E1008 00:56:23.984647 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:56:38 crc kubenswrapper[4871]: I1008 00:56:38.983671 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:56:38 crc kubenswrapper[4871]: E1008 00:56:38.985069 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:56:52 crc kubenswrapper[4871]: I1008 00:56:52.983910 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:56:52 crc kubenswrapper[4871]: E1008 00:56:52.985331 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:57:03 crc kubenswrapper[4871]: I1008 00:57:03.982358 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:57:03 crc kubenswrapper[4871]: E1008 00:57:03.983393 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.160745 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b944m/must-gather-xbv7v"] Oct 08 00:57:17 crc kubenswrapper[4871]: E1008 00:57:17.161577 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerName="registry-server" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.161588 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerName="registry-server" Oct 08 00:57:17 crc kubenswrapper[4871]: E1008 00:57:17.161608 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6206f2-ab54-4ec5-97e6-ed2460f6ee30" containerName="adoption" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.161614 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6206f2-ab54-4ec5-97e6-ed2460f6ee30" containerName="adoption" Oct 08 00:57:17 crc kubenswrapper[4871]: E1008 00:57:17.161642 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerName="extract-content" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.161648 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerName="extract-content" Oct 08 00:57:17 crc kubenswrapper[4871]: E1008 00:57:17.161659 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerName="extract-utilities" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.161664 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerName="extract-utilities" Oct 08 00:57:17 crc kubenswrapper[4871]: E1008 00:57:17.161671 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c49b518-7037-4062-99ad-072d766d43a3" containerName="adoption" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.161676 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c49b518-7037-4062-99ad-072d766d43a3" containerName="adoption" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.161889 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc3765dc-3b79-40e4-9dd2-03461978552f" containerName="registry-server" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.161907 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c49b518-7037-4062-99ad-072d766d43a3" containerName="adoption" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.161918 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b6206f2-ab54-4ec5-97e6-ed2460f6ee30" containerName="adoption" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.163023 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.167358 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b944m"/"openshift-service-ca.crt" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.167575 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b944m"/"kube-root-ca.crt" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.186558 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b944m/must-gather-xbv7v"] Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.280995 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5ab81d82-f9f2-41f2-adeb-eab439259a96-must-gather-output\") pod \"must-gather-xbv7v\" (UID: \"5ab81d82-f9f2-41f2-adeb-eab439259a96\") " pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.281101 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrljb\" (UniqueName: \"kubernetes.io/projected/5ab81d82-f9f2-41f2-adeb-eab439259a96-kube-api-access-wrljb\") pod \"must-gather-xbv7v\" (UID: \"5ab81d82-f9f2-41f2-adeb-eab439259a96\") " pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.383671 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5ab81d82-f9f2-41f2-adeb-eab439259a96-must-gather-output\") pod \"must-gather-xbv7v\" (UID: \"5ab81d82-f9f2-41f2-adeb-eab439259a96\") " pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.383720 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrljb\" (UniqueName: \"kubernetes.io/projected/5ab81d82-f9f2-41f2-adeb-eab439259a96-kube-api-access-wrljb\") pod \"must-gather-xbv7v\" (UID: \"5ab81d82-f9f2-41f2-adeb-eab439259a96\") " pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.384264 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5ab81d82-f9f2-41f2-adeb-eab439259a96-must-gather-output\") pod \"must-gather-xbv7v\" (UID: \"5ab81d82-f9f2-41f2-adeb-eab439259a96\") " pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.404612 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrljb\" (UniqueName: \"kubernetes.io/projected/5ab81d82-f9f2-41f2-adeb-eab439259a96-kube-api-access-wrljb\") pod \"must-gather-xbv7v\" (UID: \"5ab81d82-f9f2-41f2-adeb-eab439259a96\") " pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.484549 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 00:57:17 crc kubenswrapper[4871]: I1008 00:57:17.983570 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:57:17 crc kubenswrapper[4871]: E1008 00:57:17.983997 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:57:18 crc kubenswrapper[4871]: I1008 00:57:18.155396 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b944m/must-gather-xbv7v"] Oct 08 00:57:18 crc kubenswrapper[4871]: I1008 00:57:18.160366 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:57:18 crc kubenswrapper[4871]: I1008 00:57:18.268604 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/must-gather-xbv7v" event={"ID":"5ab81d82-f9f2-41f2-adeb-eab439259a96","Type":"ContainerStarted","Data":"a84788da2c82d59f4f17da83e9c5390b52b8dd285907b706ab0807d838e6e3bd"} Oct 08 00:57:23 crc kubenswrapper[4871]: I1008 00:57:23.326568 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/must-gather-xbv7v" event={"ID":"5ab81d82-f9f2-41f2-adeb-eab439259a96","Type":"ContainerStarted","Data":"b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2"} Oct 08 00:57:23 crc kubenswrapper[4871]: I1008 00:57:23.327937 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/must-gather-xbv7v" event={"ID":"5ab81d82-f9f2-41f2-adeb-eab439259a96","Type":"ContainerStarted","Data":"49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3"} Oct 08 00:57:23 crc kubenswrapper[4871]: I1008 00:57:23.347848 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b944m/must-gather-xbv7v" podStartSLOduration=1.960600516 podStartE2EDuration="6.347784939s" podCreationTimestamp="2025-10-08 00:57:17 +0000 UTC" firstStartedPulling="2025-10-08 00:57:18.16030303 +0000 UTC m=+10111.963001103" lastFinishedPulling="2025-10-08 00:57:22.547487443 +0000 UTC m=+10116.350185526" observedRunningTime="2025-10-08 00:57:23.33978622 +0000 UTC m=+10117.142484333" watchObservedRunningTime="2025-10-08 00:57:23.347784939 +0000 UTC m=+10117.150483042" Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.117619 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b944m/crc-debug-9cvsq"] Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.120162 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.121922 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-b944m"/"default-dockercfg-f47bl" Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.246722 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e9797ed-0ede-490a-ad13-a640600a5e9a-host\") pod \"crc-debug-9cvsq\" (UID: \"6e9797ed-0ede-490a-ad13-a640600a5e9a\") " pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.246870 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbt8h\" (UniqueName: \"kubernetes.io/projected/6e9797ed-0ede-490a-ad13-a640600a5e9a-kube-api-access-tbt8h\") pod \"crc-debug-9cvsq\" (UID: \"6e9797ed-0ede-490a-ad13-a640600a5e9a\") " pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.349529 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbt8h\" (UniqueName: \"kubernetes.io/projected/6e9797ed-0ede-490a-ad13-a640600a5e9a-kube-api-access-tbt8h\") pod \"crc-debug-9cvsq\" (UID: \"6e9797ed-0ede-490a-ad13-a640600a5e9a\") " pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.350139 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e9797ed-0ede-490a-ad13-a640600a5e9a-host\") pod \"crc-debug-9cvsq\" (UID: \"6e9797ed-0ede-490a-ad13-a640600a5e9a\") " pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.350227 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e9797ed-0ede-490a-ad13-a640600a5e9a-host\") pod \"crc-debug-9cvsq\" (UID: \"6e9797ed-0ede-490a-ad13-a640600a5e9a\") " pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.370202 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbt8h\" (UniqueName: \"kubernetes.io/projected/6e9797ed-0ede-490a-ad13-a640600a5e9a-kube-api-access-tbt8h\") pod \"crc-debug-9cvsq\" (UID: \"6e9797ed-0ede-490a-ad13-a640600a5e9a\") " pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:57:27 crc kubenswrapper[4871]: I1008 00:57:27.452156 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:57:28 crc kubenswrapper[4871]: I1008 00:57:28.405350 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/crc-debug-9cvsq" event={"ID":"6e9797ed-0ede-490a-ad13-a640600a5e9a","Type":"ContainerStarted","Data":"5103b6845bd6ba128ddf02302e3b42fbb5a5df213466b4c50a444fb6c0fd13df"} Oct 08 00:57:29 crc kubenswrapper[4871]: I1008 00:57:29.982499 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:57:29 crc kubenswrapper[4871]: E1008 00:57:29.983105 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:57:37 crc kubenswrapper[4871]: I1008 00:57:37.522632 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/crc-debug-9cvsq" event={"ID":"6e9797ed-0ede-490a-ad13-a640600a5e9a","Type":"ContainerStarted","Data":"c4019f5b4880889c474c9e2e7d457812e0296a8c8068daa31b5f96eef3185b3e"} Oct 08 00:57:37 crc kubenswrapper[4871]: I1008 00:57:37.542441 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b944m/crc-debug-9cvsq" podStartSLOduration=1.211415493 podStartE2EDuration="10.542421318s" podCreationTimestamp="2025-10-08 00:57:27 +0000 UTC" firstStartedPulling="2025-10-08 00:57:27.514958795 +0000 UTC m=+10121.317656868" lastFinishedPulling="2025-10-08 00:57:36.84596463 +0000 UTC m=+10130.648662693" observedRunningTime="2025-10-08 00:57:37.533892965 +0000 UTC m=+10131.336591048" watchObservedRunningTime="2025-10-08 00:57:37.542421318 +0000 UTC m=+10131.345119401" Oct 08 00:57:44 crc kubenswrapper[4871]: I1008 00:57:44.982942 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:57:44 crc kubenswrapper[4871]: E1008 00:57:44.984034 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.070827 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jh4rl"] Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.076272 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.087958 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh4rl"] Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.120013 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-utilities\") pod \"redhat-marketplace-jh4rl\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.120240 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqj6g\" (UniqueName: \"kubernetes.io/projected/539bc38f-ee93-4b12-a00c-ccd4ac10221d-kube-api-access-zqj6g\") pod \"redhat-marketplace-jh4rl\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.120284 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-catalog-content\") pod \"redhat-marketplace-jh4rl\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.222752 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqj6g\" (UniqueName: \"kubernetes.io/projected/539bc38f-ee93-4b12-a00c-ccd4ac10221d-kube-api-access-zqj6g\") pod \"redhat-marketplace-jh4rl\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.222911 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-catalog-content\") pod \"redhat-marketplace-jh4rl\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.222976 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-utilities\") pod \"redhat-marketplace-jh4rl\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.223379 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-catalog-content\") pod \"redhat-marketplace-jh4rl\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.223488 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-utilities\") pod \"redhat-marketplace-jh4rl\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.378631 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqj6g\" (UniqueName: \"kubernetes.io/projected/539bc38f-ee93-4b12-a00c-ccd4ac10221d-kube-api-access-zqj6g\") pod \"redhat-marketplace-jh4rl\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:53 crc kubenswrapper[4871]: I1008 00:57:53.437278 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.047007 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh4rl"] Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.069659 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-brt8h"] Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.075724 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.093262 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-brt8h"] Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.250584 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1e7f6c-0af7-462e-8972-3251f7b091e7-catalog-content\") pod \"certified-operators-brt8h\" (UID: \"bc1e7f6c-0af7-462e-8972-3251f7b091e7\") " pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.251023 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l4qb\" (UniqueName: \"kubernetes.io/projected/bc1e7f6c-0af7-462e-8972-3251f7b091e7-kube-api-access-7l4qb\") pod \"certified-operators-brt8h\" (UID: \"bc1e7f6c-0af7-462e-8972-3251f7b091e7\") " pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.251106 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1e7f6c-0af7-462e-8972-3251f7b091e7-utilities\") pod \"certified-operators-brt8h\" (UID: \"bc1e7f6c-0af7-462e-8972-3251f7b091e7\") " pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.352429 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l4qb\" (UniqueName: \"kubernetes.io/projected/bc1e7f6c-0af7-462e-8972-3251f7b091e7-kube-api-access-7l4qb\") pod \"certified-operators-brt8h\" (UID: \"bc1e7f6c-0af7-462e-8972-3251f7b091e7\") " pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.352541 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1e7f6c-0af7-462e-8972-3251f7b091e7-utilities\") pod \"certified-operators-brt8h\" (UID: \"bc1e7f6c-0af7-462e-8972-3251f7b091e7\") " pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.352631 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1e7f6c-0af7-462e-8972-3251f7b091e7-catalog-content\") pod \"certified-operators-brt8h\" (UID: \"bc1e7f6c-0af7-462e-8972-3251f7b091e7\") " pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.353562 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1e7f6c-0af7-462e-8972-3251f7b091e7-catalog-content\") pod \"certified-operators-brt8h\" (UID: \"bc1e7f6c-0af7-462e-8972-3251f7b091e7\") " pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.353683 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1e7f6c-0af7-462e-8972-3251f7b091e7-utilities\") pod \"certified-operators-brt8h\" (UID: \"bc1e7f6c-0af7-462e-8972-3251f7b091e7\") " pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.375224 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l4qb\" (UniqueName: \"kubernetes.io/projected/bc1e7f6c-0af7-462e-8972-3251f7b091e7-kube-api-access-7l4qb\") pod \"certified-operators-brt8h\" (UID: \"bc1e7f6c-0af7-462e-8972-3251f7b091e7\") " pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.423605 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.739951 4871 generic.go:334] "Generic (PLEG): container finished" podID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerID="819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92" exitCode=0 Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.739994 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh4rl" event={"ID":"539bc38f-ee93-4b12-a00c-ccd4ac10221d","Type":"ContainerDied","Data":"819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92"} Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.740025 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh4rl" event={"ID":"539bc38f-ee93-4b12-a00c-ccd4ac10221d","Type":"ContainerStarted","Data":"58099f82e3b091a9c5ee576280ecb1ce24249fb77b1a6d6f81626527a0afa38f"} Oct 08 00:57:54 crc kubenswrapper[4871]: I1008 00:57:54.977169 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-brt8h"] Oct 08 00:57:55 crc kubenswrapper[4871]: I1008 00:57:55.750194 4871 generic.go:334] "Generic (PLEG): container finished" podID="bc1e7f6c-0af7-462e-8972-3251f7b091e7" containerID="9e2c3b733055a7399c8f892a40952f40f6ad3953efc72a5312ad7d5ba1c78087" exitCode=0 Oct 08 00:57:55 crc kubenswrapper[4871]: I1008 00:57:55.750297 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brt8h" event={"ID":"bc1e7f6c-0af7-462e-8972-3251f7b091e7","Type":"ContainerDied","Data":"9e2c3b733055a7399c8f892a40952f40f6ad3953efc72a5312ad7d5ba1c78087"} Oct 08 00:57:55 crc kubenswrapper[4871]: I1008 00:57:55.750830 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brt8h" event={"ID":"bc1e7f6c-0af7-462e-8972-3251f7b091e7","Type":"ContainerStarted","Data":"d847b70db647dda9691862507b25ab5dbd2b0e07ab552d94159bf2efb5d6ba2f"} Oct 08 00:57:55 crc kubenswrapper[4871]: I1008 00:57:55.982340 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:57:55 crc kubenswrapper[4871]: E1008 00:57:55.982669 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 00:57:56 crc kubenswrapper[4871]: I1008 00:57:56.762351 4871 generic.go:334] "Generic (PLEG): container finished" podID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerID="d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22" exitCode=0 Oct 08 00:57:56 crc kubenswrapper[4871]: I1008 00:57:56.762439 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh4rl" event={"ID":"539bc38f-ee93-4b12-a00c-ccd4ac10221d","Type":"ContainerDied","Data":"d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22"} Oct 08 00:57:57 crc kubenswrapper[4871]: I1008 00:57:57.781744 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh4rl" event={"ID":"539bc38f-ee93-4b12-a00c-ccd4ac10221d","Type":"ContainerStarted","Data":"781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10"} Oct 08 00:57:57 crc kubenswrapper[4871]: I1008 00:57:57.805843 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jh4rl" podStartSLOduration=2.370661158 podStartE2EDuration="4.805811833s" podCreationTimestamp="2025-10-08 00:57:53 +0000 UTC" firstStartedPulling="2025-10-08 00:57:54.741980045 +0000 UTC m=+10148.544678118" lastFinishedPulling="2025-10-08 00:57:57.17713072 +0000 UTC m=+10150.979828793" observedRunningTime="2025-10-08 00:57:57.79956881 +0000 UTC m=+10151.602266883" watchObservedRunningTime="2025-10-08 00:57:57.805811833 +0000 UTC m=+10151.608509916" Oct 08 00:58:01 crc kubenswrapper[4871]: I1008 00:58:01.821720 4871 generic.go:334] "Generic (PLEG): container finished" podID="bc1e7f6c-0af7-462e-8972-3251f7b091e7" containerID="34c27c6c94d37f16cde8fcffacd9fb55d15494bcb45e655f568c40d0689c2230" exitCode=0 Oct 08 00:58:01 crc kubenswrapper[4871]: I1008 00:58:01.821809 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brt8h" event={"ID":"bc1e7f6c-0af7-462e-8972-3251f7b091e7","Type":"ContainerDied","Data":"34c27c6c94d37f16cde8fcffacd9fb55d15494bcb45e655f568c40d0689c2230"} Oct 08 00:58:03 crc kubenswrapper[4871]: I1008 00:58:03.437437 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:58:03 crc kubenswrapper[4871]: I1008 00:58:03.438036 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:58:03 crc kubenswrapper[4871]: I1008 00:58:03.525058 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:58:03 crc kubenswrapper[4871]: I1008 00:58:03.843231 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brt8h" event={"ID":"bc1e7f6c-0af7-462e-8972-3251f7b091e7","Type":"ContainerStarted","Data":"23dc54d99be07b46e93b74030ea1397582b68b764299a417516d9a12fe0920f0"} Oct 08 00:58:03 crc kubenswrapper[4871]: I1008 00:58:03.866913 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-brt8h" podStartSLOduration=3.001898913 podStartE2EDuration="9.866889297s" podCreationTimestamp="2025-10-08 00:57:54 +0000 UTC" firstStartedPulling="2025-10-08 00:57:55.752037431 +0000 UTC m=+10149.554735504" lastFinishedPulling="2025-10-08 00:58:02.617027815 +0000 UTC m=+10156.419725888" observedRunningTime="2025-10-08 00:58:03.857243505 +0000 UTC m=+10157.659941578" watchObservedRunningTime="2025-10-08 00:58:03.866889297 +0000 UTC m=+10157.669587370" Oct 08 00:58:03 crc kubenswrapper[4871]: I1008 00:58:03.895767 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:58:04 crc kubenswrapper[4871]: I1008 00:58:04.424042 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:58:04 crc kubenswrapper[4871]: I1008 00:58:04.425429 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:58:04 crc kubenswrapper[4871]: I1008 00:58:04.613179 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:58:04 crc kubenswrapper[4871]: I1008 00:58:04.771949 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh4rl"] Oct 08 00:58:05 crc kubenswrapper[4871]: I1008 00:58:05.865137 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jh4rl" podUID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerName="registry-server" containerID="cri-o://781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10" gracePeriod=2 Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.403262 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.538864 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-catalog-content\") pod \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.538971 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-utilities\") pod \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.539015 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqj6g\" (UniqueName: \"kubernetes.io/projected/539bc38f-ee93-4b12-a00c-ccd4ac10221d-kube-api-access-zqj6g\") pod \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\" (UID: \"539bc38f-ee93-4b12-a00c-ccd4ac10221d\") " Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.539936 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-utilities" (OuterVolumeSpecName: "utilities") pod "539bc38f-ee93-4b12-a00c-ccd4ac10221d" (UID: "539bc38f-ee93-4b12-a00c-ccd4ac10221d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.548968 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/539bc38f-ee93-4b12-a00c-ccd4ac10221d-kube-api-access-zqj6g" (OuterVolumeSpecName: "kube-api-access-zqj6g") pod "539bc38f-ee93-4b12-a00c-ccd4ac10221d" (UID: "539bc38f-ee93-4b12-a00c-ccd4ac10221d"). InnerVolumeSpecName "kube-api-access-zqj6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.553358 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "539bc38f-ee93-4b12-a00c-ccd4ac10221d" (UID: "539bc38f-ee93-4b12-a00c-ccd4ac10221d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.642046 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.642441 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/539bc38f-ee93-4b12-a00c-ccd4ac10221d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.642461 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqj6g\" (UniqueName: \"kubernetes.io/projected/539bc38f-ee93-4b12-a00c-ccd4ac10221d-kube-api-access-zqj6g\") on node \"crc\" DevicePath \"\"" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.885312 4871 generic.go:334] "Generic (PLEG): container finished" podID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerID="781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10" exitCode=0 Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.885373 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh4rl" event={"ID":"539bc38f-ee93-4b12-a00c-ccd4ac10221d","Type":"ContainerDied","Data":"781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10"} Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.885407 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh4rl" event={"ID":"539bc38f-ee93-4b12-a00c-ccd4ac10221d","Type":"ContainerDied","Data":"58099f82e3b091a9c5ee576280ecb1ce24249fb77b1a6d6f81626527a0afa38f"} Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.885428 4871 scope.go:117] "RemoveContainer" containerID="781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.885620 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jh4rl" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.930117 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh4rl"] Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.931763 4871 scope.go:117] "RemoveContainer" containerID="d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.944274 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh4rl"] Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.977858 4871 scope.go:117] "RemoveContainer" containerID="819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92" Oct 08 00:58:06 crc kubenswrapper[4871]: I1008 00:58:06.982994 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 00:58:07 crc kubenswrapper[4871]: I1008 00:58:07.008151 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" path="/var/lib/kubelet/pods/539bc38f-ee93-4b12-a00c-ccd4ac10221d/volumes" Oct 08 00:58:07 crc kubenswrapper[4871]: I1008 00:58:07.023760 4871 scope.go:117] "RemoveContainer" containerID="781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10" Oct 08 00:58:07 crc kubenswrapper[4871]: E1008 00:58:07.024170 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10\": container with ID starting with 781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10 not found: ID does not exist" containerID="781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10" Oct 08 00:58:07 crc kubenswrapper[4871]: I1008 00:58:07.024214 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10"} err="failed to get container status \"781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10\": rpc error: code = NotFound desc = could not find container \"781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10\": container with ID starting with 781a4e5ee231a5bed726473635467ba9d83a88d99be1549995607a5f25337c10 not found: ID does not exist" Oct 08 00:58:07 crc kubenswrapper[4871]: I1008 00:58:07.024239 4871 scope.go:117] "RemoveContainer" containerID="d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22" Oct 08 00:58:07 crc kubenswrapper[4871]: E1008 00:58:07.024474 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22\": container with ID starting with d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22 not found: ID does not exist" containerID="d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22" Oct 08 00:58:07 crc kubenswrapper[4871]: I1008 00:58:07.024504 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22"} err="failed to get container status \"d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22\": rpc error: code = NotFound desc = could not find container \"d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22\": container with ID starting with d83cf2155baef60074e3016a43bef62ea55df9500b4730423155a6c826cb0e22 not found: ID does not exist" Oct 08 00:58:07 crc kubenswrapper[4871]: I1008 00:58:07.024518 4871 scope.go:117] "RemoveContainer" containerID="819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92" Oct 08 00:58:07 crc kubenswrapper[4871]: E1008 00:58:07.024706 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92\": container with ID starting with 819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92 not found: ID does not exist" containerID="819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92" Oct 08 00:58:07 crc kubenswrapper[4871]: I1008 00:58:07.024726 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92"} err="failed to get container status \"819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92\": rpc error: code = NotFound desc = could not find container \"819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92\": container with ID starting with 819541e9192af4ec9f8dfec25bf35056e2eda337d5647d3061600f1af743ca92 not found: ID does not exist" Oct 08 00:58:07 crc kubenswrapper[4871]: I1008 00:58:07.907516 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"6b5bdcf7675888de17bfd75524c8b30d3e8713d1d2a3c5cd0e1017c4ecc59890"} Oct 08 00:58:14 crc kubenswrapper[4871]: I1008 00:58:14.482590 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-brt8h" Oct 08 00:58:14 crc kubenswrapper[4871]: I1008 00:58:14.554501 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-brt8h"] Oct 08 00:58:14 crc kubenswrapper[4871]: I1008 00:58:14.602321 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8zcc7"] Oct 08 00:58:14 crc kubenswrapper[4871]: I1008 00:58:14.602582 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8zcc7" podUID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerName="registry-server" containerID="cri-o://6808860fccb52642eb89c2255d10654e90788a19449d81fe8a661d7fd1b716b8" gracePeriod=2 Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:14.999568 4871 generic.go:334] "Generic (PLEG): container finished" podID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerID="6808860fccb52642eb89c2255d10654e90788a19449d81fe8a661d7fd1b716b8" exitCode=0 Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.001363 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8zcc7" event={"ID":"5f8d6dfe-9200-4130-ba88-34934aa55ea0","Type":"ContainerDied","Data":"6808860fccb52642eb89c2255d10654e90788a19449d81fe8a661d7fd1b716b8"} Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.299409 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.440190 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-catalog-content\") pod \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.440373 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-utilities\") pod \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.440429 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqgsf\" (UniqueName: \"kubernetes.io/projected/5f8d6dfe-9200-4130-ba88-34934aa55ea0-kube-api-access-bqgsf\") pod \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\" (UID: \"5f8d6dfe-9200-4130-ba88-34934aa55ea0\") " Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.441788 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-utilities" (OuterVolumeSpecName: "utilities") pod "5f8d6dfe-9200-4130-ba88-34934aa55ea0" (UID: "5f8d6dfe-9200-4130-ba88-34934aa55ea0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.450119 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f8d6dfe-9200-4130-ba88-34934aa55ea0-kube-api-access-bqgsf" (OuterVolumeSpecName: "kube-api-access-bqgsf") pod "5f8d6dfe-9200-4130-ba88-34934aa55ea0" (UID: "5f8d6dfe-9200-4130-ba88-34934aa55ea0"). InnerVolumeSpecName "kube-api-access-bqgsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.533040 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f8d6dfe-9200-4130-ba88-34934aa55ea0" (UID: "5f8d6dfe-9200-4130-ba88-34934aa55ea0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.542703 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.542750 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8d6dfe-9200-4130-ba88-34934aa55ea0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:58:15 crc kubenswrapper[4871]: I1008 00:58:15.542761 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqgsf\" (UniqueName: \"kubernetes.io/projected/5f8d6dfe-9200-4130-ba88-34934aa55ea0-kube-api-access-bqgsf\") on node \"crc\" DevicePath \"\"" Oct 08 00:58:16 crc kubenswrapper[4871]: I1008 00:58:16.016984 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8zcc7" event={"ID":"5f8d6dfe-9200-4130-ba88-34934aa55ea0","Type":"ContainerDied","Data":"d12d78894994ac3f65cc51a562b23aff9ad8b7354ec74367fdaeede577c76dfe"} Oct 08 00:58:16 crc kubenswrapper[4871]: I1008 00:58:16.017326 4871 scope.go:117] "RemoveContainer" containerID="6808860fccb52642eb89c2255d10654e90788a19449d81fe8a661d7fd1b716b8" Oct 08 00:58:16 crc kubenswrapper[4871]: I1008 00:58:16.017037 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8zcc7" Oct 08 00:58:16 crc kubenswrapper[4871]: I1008 00:58:16.042814 4871 scope.go:117] "RemoveContainer" containerID="a492cd44e1f1c4acdcb13e6cb9e34082430ff9e486e71aed68d265c112683042" Oct 08 00:58:16 crc kubenswrapper[4871]: I1008 00:58:16.053225 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8zcc7"] Oct 08 00:58:16 crc kubenswrapper[4871]: I1008 00:58:16.062363 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8zcc7"] Oct 08 00:58:16 crc kubenswrapper[4871]: I1008 00:58:16.103285 4871 scope.go:117] "RemoveContainer" containerID="9b6989ced397e6a745981a1ccfdedb5d4a15505384edb280e8ff32c7596e551e" Oct 08 00:58:17 crc kubenswrapper[4871]: I1008 00:58:17.008058 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" path="/var/lib/kubelet/pods/5f8d6dfe-9200-4130-ba88-34934aa55ea0/volumes" Oct 08 00:58:56 crc kubenswrapper[4871]: I1008 00:58:56.207021 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ed4402de-a3a0-4c54-9675-a0766947a554/init-config-reloader/0.log" Oct 08 00:58:56 crc kubenswrapper[4871]: I1008 00:58:56.412958 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ed4402de-a3a0-4c54-9675-a0766947a554/init-config-reloader/0.log" Oct 08 00:58:56 crc kubenswrapper[4871]: I1008 00:58:56.443682 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ed4402de-a3a0-4c54-9675-a0766947a554/alertmanager/0.log" Oct 08 00:58:56 crc kubenswrapper[4871]: I1008 00:58:56.594947 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ed4402de-a3a0-4c54-9675-a0766947a554/config-reloader/0.log" Oct 08 00:58:56 crc kubenswrapper[4871]: I1008 00:58:56.786100 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ed1a54c1-15bb-427b-af24-0fb7909985af/aodh-api/0.log" Oct 08 00:58:57 crc kubenswrapper[4871]: I1008 00:58:57.131885 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ed1a54c1-15bb-427b-af24-0fb7909985af/aodh-evaluator/0.log" Oct 08 00:58:57 crc kubenswrapper[4871]: I1008 00:58:57.175760 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ed1a54c1-15bb-427b-af24-0fb7909985af/aodh-listener/0.log" Oct 08 00:58:57 crc kubenswrapper[4871]: I1008 00:58:57.354829 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ed1a54c1-15bb-427b-af24-0fb7909985af/aodh-notifier/0.log" Oct 08 00:58:57 crc kubenswrapper[4871]: I1008 00:58:57.525632 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6849865b9b-99rd5_61ba6fb5-5888-410b-938a-dc06179dd26f/barbican-api/0.log" Oct 08 00:58:57 crc kubenswrapper[4871]: I1008 00:58:57.632281 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6849865b9b-99rd5_61ba6fb5-5888-410b-938a-dc06179dd26f/barbican-api-log/0.log" Oct 08 00:58:57 crc kubenswrapper[4871]: I1008 00:58:57.862550 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-74dc8ccf74-5tktj_e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7/barbican-keystone-listener/0.log" Oct 08 00:58:58 crc kubenswrapper[4871]: I1008 00:58:58.051187 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-74dc8ccf74-5tktj_e2c0ce1f-9c95-41e8-b4b8-90a4a604cfb7/barbican-keystone-listener-log/0.log" Oct 08 00:58:58 crc kubenswrapper[4871]: I1008 00:58:58.230380 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5cc676b4f5-92cx4_f2cf443f-a41e-4c1a-a488-0a90f44b8e74/barbican-worker/0.log" Oct 08 00:58:58 crc kubenswrapper[4871]: I1008 00:58:58.236550 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5cc676b4f5-92cx4_f2cf443f-a41e-4c1a-a488-0a90f44b8e74/barbican-worker-log/0.log" Oct 08 00:58:58 crc kubenswrapper[4871]: I1008 00:58:58.456288 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-xxtff_c5eb167d-5c30-4354-888c-25c58fdfd2e9/bootstrap-openstack-openstack-cell1/0.log" Oct 08 00:58:58 crc kubenswrapper[4871]: I1008 00:58:58.569180 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_304c982b-0677-407f-88c4-4efbbc47a363/ceilometer-central-agent/0.log" Oct 08 00:58:58 crc kubenswrapper[4871]: I1008 00:58:58.704595 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_304c982b-0677-407f-88c4-4efbbc47a363/proxy-httpd/0.log" Oct 08 00:58:58 crc kubenswrapper[4871]: I1008 00:58:58.705782 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_304c982b-0677-407f-88c4-4efbbc47a363/ceilometer-notification-agent/0.log" Oct 08 00:58:58 crc kubenswrapper[4871]: I1008 00:58:58.739665 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_304c982b-0677-407f-88c4-4efbbc47a363/sg-core/0.log" Oct 08 00:58:58 crc kubenswrapper[4871]: I1008 00:58:58.885260 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-j98xd_04f0cc94-5c7f-4cfd-94ea-f0e9556f4cc6/ceph-client-openstack-openstack-cell1/0.log" Oct 08 00:58:59 crc kubenswrapper[4871]: I1008 00:58:59.185218 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_bf095ee8-6cee-4ab7-888d-8cea407f2e82/cinder-api/0.log" Oct 08 00:58:59 crc kubenswrapper[4871]: I1008 00:58:59.214700 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_bf095ee8-6cee-4ab7-888d-8cea407f2e82/cinder-api-log/0.log" Oct 08 00:58:59 crc kubenswrapper[4871]: I1008 00:58:59.455146 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_06fc942e-87b3-435b-97c3-4a42d7d78abe/probe/0.log" Oct 08 00:58:59 crc kubenswrapper[4871]: I1008 00:58:59.518664 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_06fc942e-87b3-435b-97c3-4a42d7d78abe/cinder-backup/0.log" Oct 08 00:58:59 crc kubenswrapper[4871]: I1008 00:58:59.643284 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-db-purge-29331361-xd69z_6dacc481-da68-45bf-b7a2-24f796328ee1/cinder-db-purge/0.log" Oct 08 00:58:59 crc kubenswrapper[4871]: I1008 00:58:59.743090 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_213171d1-7fd7-4b97-84cb-7917496a94a4/cinder-scheduler/0.log" Oct 08 00:58:59 crc kubenswrapper[4871]: I1008 00:58:59.873974 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_213171d1-7fd7-4b97-84cb-7917496a94a4/probe/0.log" Oct 08 00:59:00 crc kubenswrapper[4871]: I1008 00:59:00.018439 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_60ebc4f7-1bb0-481e-871e-a36c56c744a2/cinder-volume/0.log" Oct 08 00:59:00 crc kubenswrapper[4871]: I1008 00:59:00.168283 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_60ebc4f7-1bb0-481e-871e-a36c56c744a2/probe/0.log" Oct 08 00:59:00 crc kubenswrapper[4871]: I1008 00:59:00.229037 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-4przw_f6c907cd-13ef-4d90-be16-aeeb560c2d9e/configure-network-openstack-openstack-cell1/0.log" Oct 08 00:59:00 crc kubenswrapper[4871]: I1008 00:59:00.502470 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-t9g5h_b587d525-db27-497c-9539-f5492c8e4b33/configure-os-openstack-openstack-cell1/0.log" Oct 08 00:59:00 crc kubenswrapper[4871]: I1008 00:59:00.676621 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-855fbfb687-7pd4r_ad432e67-ef64-408a-9b6b-4b76a60d308e/init/0.log" Oct 08 00:59:00 crc kubenswrapper[4871]: I1008 00:59:00.881158 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-855fbfb687-7pd4r_ad432e67-ef64-408a-9b6b-4b76a60d308e/init/0.log" Oct 08 00:59:00 crc kubenswrapper[4871]: I1008 00:59:00.917240 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-855fbfb687-7pd4r_ad432e67-ef64-408a-9b6b-4b76a60d308e/dnsmasq-dns/0.log" Oct 08 00:59:01 crc kubenswrapper[4871]: I1008 00:59:01.867626 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-db-purge-29331361-shv9p_80141182-14e1-4535-a176-4c69d98eaaa5/glance-dbpurge/0.log" Oct 08 00:59:01 crc kubenswrapper[4871]: I1008 00:59:01.884171 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-td4fb_0a1b4a8b-4d71-4bd5-b9dc-e784713c2793/download-cache-openstack-openstack-cell1/0.log" Oct 08 00:59:02 crc kubenswrapper[4871]: I1008 00:59:02.098323 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_02e76f14-d4ce-486a-8979-ff8a12871783/glance-httpd/0.log" Oct 08 00:59:02 crc kubenswrapper[4871]: I1008 00:59:02.124813 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_02e76f14-d4ce-486a-8979-ff8a12871783/glance-log/0.log" Oct 08 00:59:02 crc kubenswrapper[4871]: I1008 00:59:02.362431 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c54d4124-f070-4e87-bfd7-6f319c374d36/glance-log/0.log" Oct 08 00:59:02 crc kubenswrapper[4871]: I1008 00:59:02.365997 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c54d4124-f070-4e87-bfd7-6f319c374d36/glance-httpd/0.log" Oct 08 00:59:02 crc kubenswrapper[4871]: I1008 00:59:02.647524 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5b8c8579fc-wmrqt_0193e1f0-49ff-4945-a039-a31e3431dec3/heat-api/0.log" Oct 08 00:59:02 crc kubenswrapper[4871]: I1008 00:59:02.950416 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5c75875cf4-v2bhz_c0bd4d03-7aba-460f-84a0-1fa48a81e822/heat-cfnapi/0.log" Oct 08 00:59:03 crc kubenswrapper[4871]: I1008 00:59:03.629922 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-574ddff48-pw8c9_e92c3a0b-ecce-4129-9cb4-d6bf8f597225/heat-engine/0.log" Oct 08 00:59:03 crc kubenswrapper[4871]: I1008 00:59:03.715041 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f455d5855-8bgr8_674d0e65-ecd4-4b3a-a734-4c01e2b4192b/horizon/0.log" Oct 08 00:59:03 crc kubenswrapper[4871]: I1008 00:59:03.916619 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f455d5855-8bgr8_674d0e65-ecd4-4b3a-a734-4c01e2b4192b/horizon-log/0.log" Oct 08 00:59:03 crc kubenswrapper[4871]: I1008 00:59:03.984388 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-tsvxx_24161d8f-3e5c-4225-aecf-89a7d3d825cc/install-certs-openstack-openstack-cell1/0.log" Oct 08 00:59:04 crc kubenswrapper[4871]: I1008 00:59:04.139313 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-pthdq_1e32fa4d-959c-4011-8b54-aa7d820ff228/install-os-openstack-openstack-cell1/0.log" Oct 08 00:59:04 crc kubenswrapper[4871]: I1008 00:59:04.305526 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29331361-9wnxt_35542928-206a-4b3f-bf2e-ba4ea4449cc3/keystone-cron/0.log" Oct 08 00:59:04 crc kubenswrapper[4871]: I1008 00:59:04.577627 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-dbc89c896-7kvkt_3500a587-fccc-460c-9848-520a06582186/keystone-api/0.log" Oct 08 00:59:04 crc kubenswrapper[4871]: I1008 00:59:04.603923 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_96e5e7ae-f05f-4645-bd64-d51333e8bc4e/kube-state-metrics/0.log" Oct 08 00:59:04 crc kubenswrapper[4871]: I1008 00:59:04.809731 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-gb8fk_328f608d-5164-4165-9cfc-76a5b1caa79c/libvirt-openstack-openstack-cell1/0.log" Oct 08 00:59:05 crc kubenswrapper[4871]: I1008 00:59:05.030868 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_95791187-6a85-4955-997d-920afa9d8adf/manila-api-log/0.log" Oct 08 00:59:05 crc kubenswrapper[4871]: I1008 00:59:05.092809 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_95791187-6a85-4955-997d-920afa9d8adf/manila-api/0.log" Oct 08 00:59:05 crc kubenswrapper[4871]: I1008 00:59:05.236079 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-db-purge-29331361-crl8s_401dc5ea-abe5-43e0-9d3f-095574675437/manila-db-purge/0.log" Oct 08 00:59:05 crc kubenswrapper[4871]: I1008 00:59:05.384076 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_e21ff4e8-d971-4e36-bdb5-ea2e5364326e/manila-scheduler/0.log" Oct 08 00:59:05 crc kubenswrapper[4871]: I1008 00:59:05.478226 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_e21ff4e8-d971-4e36-bdb5-ea2e5364326e/probe/0.log" Oct 08 00:59:05 crc kubenswrapper[4871]: I1008 00:59:05.585700 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_44e3376b-cce2-4e65-a494-c17253d07c09/manila-share/0.log" Oct 08 00:59:05 crc kubenswrapper[4871]: I1008 00:59:05.687306 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_44e3376b-cce2-4e65-a494-c17253d07c09/probe/0.log" Oct 08 00:59:06 crc kubenswrapper[4871]: I1008 00:59:06.177329 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-65f4496dcc-6mkfj_f242653b-4449-452c-a438-2fbdd2a87dba/neutron-api/0.log" Oct 08 00:59:06 crc kubenswrapper[4871]: I1008 00:59:06.511840 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-65f4496dcc-6mkfj_f242653b-4449-452c-a438-2fbdd2a87dba/neutron-httpd/0.log" Oct 08 00:59:06 crc kubenswrapper[4871]: I1008 00:59:06.787222 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-rfmrs_ac814cee-d737-4c1f-b486-93d99024c197/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 08 00:59:07 crc kubenswrapper[4871]: I1008 00:59:07.106343 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-qnn9h_e2f9a2a7-995e-46a7-9272-aa54c93afa60/neutron-metadata-openstack-openstack-cell1/0.log" Oct 08 00:59:07 crc kubenswrapper[4871]: I1008 00:59:07.416592 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-s9zww_59d68a40-edd5-4599-a9dc-5fa134a4a05c/neutron-sriov-openstack-openstack-cell1/0.log" Oct 08 00:59:07 crc kubenswrapper[4871]: I1008 00:59:07.717540 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d9f07d2a-d7a7-46b1-9e36-faff79d08add/nova-api-api/0.log" Oct 08 00:59:08 crc kubenswrapper[4871]: I1008 00:59:08.053727 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d9f07d2a-d7a7-46b1-9e36-faff79d08add/nova-api-log/0.log" Oct 08 00:59:08 crc kubenswrapper[4871]: I1008 00:59:08.322991 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_0e5e4de2-9acb-4879-bba9-7466a6b77ba2/nova-cell0-conductor-conductor/0.log" Oct 08 00:59:08 crc kubenswrapper[4871]: I1008 00:59:08.541921 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-db-purge-29331360-6zsn2_8dbba49b-2784-4094-8fa7-2c7ba25d6c0c/nova-manage/0.log" Oct 08 00:59:08 crc kubenswrapper[4871]: I1008 00:59:08.635689 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2985dd37-b0b5-4d84-9a26-cddfaf27be64/memcached/0.log" Oct 08 00:59:08 crc kubenswrapper[4871]: I1008 00:59:08.793564 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_57cb8659-97fb-41d1-b2cf-044b95643dd2/nova-cell1-conductor-conductor/0.log" Oct 08 00:59:08 crc kubenswrapper[4871]: I1008 00:59:08.846169 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-db-purge-29331360-d65r4_dcde4412-ee28-45eb-91f1-2718fa9fe0a5/nova-manage/0.log" Oct 08 00:59:09 crc kubenswrapper[4871]: I1008 00:59:09.160854 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c5ba2fb5-0f9f-41fa-816b-1db49317eb3e/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 00:59:09 crc kubenswrapper[4871]: I1008 00:59:09.403998 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldzgfk_b8b25c96-823b-433c-8512-cf364233ebe0/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 08 00:59:09 crc kubenswrapper[4871]: I1008 00:59:09.535926 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-k2kcq_bf1b098f-7fd8-4c0d-8c4e-88abab628cbe/nova-cell1-openstack-openstack-cell1/0.log" Oct 08 00:59:09 crc kubenswrapper[4871]: I1008 00:59:09.660698 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4e277fae-f0f9-469b-87da-b7c88f40871f/nova-metadata-log/0.log" Oct 08 00:59:09 crc kubenswrapper[4871]: I1008 00:59:09.818274 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4e277fae-f0f9-469b-87da-b7c88f40871f/nova-metadata-metadata/0.log" Oct 08 00:59:09 crc kubenswrapper[4871]: I1008 00:59:09.927765 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e23fd31e-e2d3-4dfe-be60-e7497b9b6748/nova-scheduler-scheduler/0.log" Oct 08 00:59:10 crc kubenswrapper[4871]: I1008 00:59:10.328862 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7bb5c546c4-ms9lr_446688c4-a538-417b-97d5-a8c2b9203f3c/init/0.log" Oct 08 00:59:10 crc kubenswrapper[4871]: I1008 00:59:10.554105 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7bb5c546c4-ms9lr_446688c4-a538-417b-97d5-a8c2b9203f3c/init/0.log" Oct 08 00:59:10 crc kubenswrapper[4871]: I1008 00:59:10.630765 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7bb5c546c4-ms9lr_446688c4-a538-417b-97d5-a8c2b9203f3c/octavia-api-provider-agent/0.log" Oct 08 00:59:10 crc kubenswrapper[4871]: I1008 00:59:10.706507 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7bb5c546c4-ms9lr_446688c4-a538-417b-97d5-a8c2b9203f3c/octavia-api/0.log" Oct 08 00:59:10 crc kubenswrapper[4871]: I1008 00:59:10.848980 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-knthn_d730e692-7f26-4844-a01d-f79e38ced5fc/init/0.log" Oct 08 00:59:11 crc kubenswrapper[4871]: I1008 00:59:11.067115 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-knthn_d730e692-7f26-4844-a01d-f79e38ced5fc/init/0.log" Oct 08 00:59:11 crc kubenswrapper[4871]: I1008 00:59:11.184867 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-knthn_d730e692-7f26-4844-a01d-f79e38ced5fc/octavia-healthmanager/0.log" Oct 08 00:59:11 crc kubenswrapper[4871]: I1008 00:59:11.270072 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-f2v96_f01a2153-7671-41bf-9c79-7bf208956897/init/0.log" Oct 08 00:59:12 crc kubenswrapper[4871]: I1008 00:59:12.223349 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-f2v96_f01a2153-7671-41bf-9c79-7bf208956897/init/0.log" Oct 08 00:59:12 crc kubenswrapper[4871]: I1008 00:59:12.252668 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-f2v96_f01a2153-7671-41bf-9c79-7bf208956897/octavia-housekeeping/0.log" Oct 08 00:59:12 crc kubenswrapper[4871]: I1008 00:59:12.440150 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-ckzbv_13bcfd35-b654-40e4-88d8-fc290ba96b14/init/0.log" Oct 08 00:59:12 crc kubenswrapper[4871]: I1008 00:59:12.593704 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-ckzbv_13bcfd35-b654-40e4-88d8-fc290ba96b14/init/0.log" Oct 08 00:59:12 crc kubenswrapper[4871]: I1008 00:59:12.601620 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-ckzbv_13bcfd35-b654-40e4-88d8-fc290ba96b14/octavia-rsyslog/0.log" Oct 08 00:59:12 crc kubenswrapper[4871]: I1008 00:59:12.822335 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-jnzlv_8eeebe46-f544-4884-b33b-06ed31f5c6f7/init/0.log" Oct 08 00:59:13 crc kubenswrapper[4871]: I1008 00:59:13.000811 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-jnzlv_8eeebe46-f544-4884-b33b-06ed31f5c6f7/init/0.log" Oct 08 00:59:13 crc kubenswrapper[4871]: I1008 00:59:13.119171 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-jnzlv_8eeebe46-f544-4884-b33b-06ed31f5c6f7/octavia-worker/0.log" Oct 08 00:59:13 crc kubenswrapper[4871]: I1008 00:59:13.216344 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f780ea63-c7a1-4fdb-828d-bf2c1492e300/mysql-bootstrap/0.log" Oct 08 00:59:14 crc kubenswrapper[4871]: I1008 00:59:14.219159 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f780ea63-c7a1-4fdb-828d-bf2c1492e300/mysql-bootstrap/0.log" Oct 08 00:59:14 crc kubenswrapper[4871]: I1008 00:59:14.256403 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f780ea63-c7a1-4fdb-828d-bf2c1492e300/galera/0.log" Oct 08 00:59:14 crc kubenswrapper[4871]: I1008 00:59:14.610127 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_08850732-6460-4677-a906-b8d932f2f1ec/mysql-bootstrap/0.log" Oct 08 00:59:14 crc kubenswrapper[4871]: I1008 00:59:14.845538 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_08850732-6460-4677-a906-b8d932f2f1ec/mysql-bootstrap/0.log" Oct 08 00:59:14 crc kubenswrapper[4871]: I1008 00:59:14.897838 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_08850732-6460-4677-a906-b8d932f2f1ec/galera/0.log" Oct 08 00:59:15 crc kubenswrapper[4871]: I1008 00:59:15.082737 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_64093265-fc84-469c-bbc0-55b87680c9e9/openstackclient/0.log" Oct 08 00:59:15 crc kubenswrapper[4871]: I1008 00:59:15.143923 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lcsmb_2c553ffe-f347-4370-a2c5-c80ac5686859/openstack-network-exporter/0.log" Oct 08 00:59:15 crc kubenswrapper[4871]: I1008 00:59:15.300948 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-8dvm9_58f6bc11-48a7-48c8-a28b-5e02c8d284c5/ovsdb-server-init/0.log" Oct 08 00:59:15 crc kubenswrapper[4871]: I1008 00:59:15.469589 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-8dvm9_58f6bc11-48a7-48c8-a28b-5e02c8d284c5/ovs-vswitchd/0.log" Oct 08 00:59:15 crc kubenswrapper[4871]: I1008 00:59:15.489165 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-8dvm9_58f6bc11-48a7-48c8-a28b-5e02c8d284c5/ovsdb-server/0.log" Oct 08 00:59:15 crc kubenswrapper[4871]: I1008 00:59:15.491494 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-8dvm9_58f6bc11-48a7-48c8-a28b-5e02c8d284c5/ovsdb-server-init/0.log" Oct 08 00:59:15 crc kubenswrapper[4871]: I1008 00:59:15.670759 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-w7c98_c38bb4f6-50ea-4b50-97f3-13a6838c4aae/ovn-controller/0.log" Oct 08 00:59:15 crc kubenswrapper[4871]: I1008 00:59:15.823461 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_df2f59e7-06b2-42a4-bd21-d972ad2f89d5/openstack-network-exporter/0.log" Oct 08 00:59:15 crc kubenswrapper[4871]: I1008 00:59:15.899708 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_df2f59e7-06b2-42a4-bd21-d972ad2f89d5/ovn-northd/0.log" Oct 08 00:59:16 crc kubenswrapper[4871]: I1008 00:59:16.046028 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-26h6r_9efec3ad-9ec3-4008-88b7-fb1c505d459b/ovn-openstack-openstack-cell1/0.log" Oct 08 00:59:16 crc kubenswrapper[4871]: I1008 00:59:16.265018 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_258f0b30-db8c-4b1f-93e1-42df4430228c/openstack-network-exporter/0.log" Oct 08 00:59:16 crc kubenswrapper[4871]: I1008 00:59:16.307651 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_258f0b30-db8c-4b1f-93e1-42df4430228c/ovsdbserver-nb/0.log" Oct 08 00:59:16 crc kubenswrapper[4871]: I1008 00:59:16.477290 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_54438dbf-34ef-4626-acb1-8c0f4db20a88/openstack-network-exporter/0.log" Oct 08 00:59:16 crc kubenswrapper[4871]: I1008 00:59:16.497777 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_54438dbf-34ef-4626-acb1-8c0f4db20a88/ovsdbserver-nb/0.log" Oct 08 00:59:16 crc kubenswrapper[4871]: I1008 00:59:16.660563 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_eda861cc-b8ae-4a99-b651-a2f31601a006/openstack-network-exporter/0.log" Oct 08 00:59:16 crc kubenswrapper[4871]: I1008 00:59:16.754508 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_eda861cc-b8ae-4a99-b651-a2f31601a006/ovsdbserver-nb/0.log" Oct 08 00:59:16 crc kubenswrapper[4871]: I1008 00:59:16.938285 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_faf833c3-7d4d-4c64-b466-3c48ef1a42c9/openstack-network-exporter/0.log" Oct 08 00:59:16 crc kubenswrapper[4871]: I1008 00:59:16.983589 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_faf833c3-7d4d-4c64-b466-3c48ef1a42c9/ovsdbserver-sb/0.log" Oct 08 00:59:17 crc kubenswrapper[4871]: I1008 00:59:17.158302 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_d38b4227-32df-4361-b559-bd3c7ef7e741/openstack-network-exporter/0.log" Oct 08 00:59:17 crc kubenswrapper[4871]: I1008 00:59:17.204169 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_d38b4227-32df-4361-b559-bd3c7ef7e741/ovsdbserver-sb/0.log" Oct 08 00:59:17 crc kubenswrapper[4871]: I1008 00:59:17.406125 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_f2d065bc-0f6a-4e23-8ffe-b9d07d512d08/openstack-network-exporter/0.log" Oct 08 00:59:17 crc kubenswrapper[4871]: I1008 00:59:17.413341 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_f2d065bc-0f6a-4e23-8ffe-b9d07d512d08/ovsdbserver-sb/0.log" Oct 08 00:59:17 crc kubenswrapper[4871]: I1008 00:59:17.645723 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-c884b686b-r8slm_019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57/placement-api/0.log" Oct 08 00:59:17 crc kubenswrapper[4871]: I1008 00:59:17.724065 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-c884b686b-r8slm_019ce7b1-d8f3-4069-85ff-3fd4bd4dcf57/placement-log/0.log" Oct 08 00:59:17 crc kubenswrapper[4871]: I1008 00:59:17.869348 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cpvxpr_be4cf803-f602-46b3-90ba-e17ed62e248c/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.060288 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9edd958f-36cf-453d-a0f8-c5f27cdfdc07/init-config-reloader/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.223667 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9edd958f-36cf-453d-a0f8-c5f27cdfdc07/config-reloader/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.231693 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9edd958f-36cf-453d-a0f8-c5f27cdfdc07/init-config-reloader/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.239114 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9edd958f-36cf-453d-a0f8-c5f27cdfdc07/prometheus/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.415109 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_50bfcc25-71f9-46ee-b7ab-7be7d920c19f/setup-container/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.416532 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9edd958f-36cf-453d-a0f8-c5f27cdfdc07/thanos-sidecar/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.589754 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_50bfcc25-71f9-46ee-b7ab-7be7d920c19f/setup-container/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.629715 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_50bfcc25-71f9-46ee-b7ab-7be7d920c19f/rabbitmq/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.806140 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2b9bb1f5-e47a-48af-9a8b-4b766881f56f/setup-container/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.966309 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2b9bb1f5-e47a-48af-9a8b-4b766881f56f/setup-container/0.log" Oct 08 00:59:18 crc kubenswrapper[4871]: I1008 00:59:18.985913 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2b9bb1f5-e47a-48af-9a8b-4b766881f56f/rabbitmq/0.log" Oct 08 00:59:19 crc kubenswrapper[4871]: I1008 00:59:19.151506 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-mj2bz_88906f81-9b44-4372-9962-6291c10fdb57/reboot-os-openstack-openstack-cell1/0.log" Oct 08 00:59:19 crc kubenswrapper[4871]: I1008 00:59:19.288522 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-j8zp6_f642efbd-d4dc-4f35-93f0-b9aa5ee9b73c/run-os-openstack-openstack-cell1/0.log" Oct 08 00:59:19 crc kubenswrapper[4871]: I1008 00:59:19.444067 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-9w8j5_d5b4e059-c3a7-41ee-a76f-104e6cfce8a0/ssh-known-hosts-openstack/0.log" Oct 08 00:59:19 crc kubenswrapper[4871]: I1008 00:59:19.645988 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-jmt5b_7b13172c-c2ed-41f2-ab24-538418c2e1b3/telemetry-openstack-openstack-cell1/0.log" Oct 08 00:59:19 crc kubenswrapper[4871]: I1008 00:59:19.786311 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-kq5z6_8a39a36c-8c17-4049-99a0-3ef681c082b3/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 08 00:59:19 crc kubenswrapper[4871]: I1008 00:59:19.881457 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-96fss_404d74ce-e450-47f2-acad-54b000cae498/validate-network-openstack-openstack-cell1/0.log" Oct 08 00:59:46 crc kubenswrapper[4871]: I1008 00:59:46.968929 4871 generic.go:334] "Generic (PLEG): container finished" podID="6e9797ed-0ede-490a-ad13-a640600a5e9a" containerID="c4019f5b4880889c474c9e2e7d457812e0296a8c8068daa31b5f96eef3185b3e" exitCode=0 Oct 08 00:59:46 crc kubenswrapper[4871]: I1008 00:59:46.969025 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/crc-debug-9cvsq" event={"ID":"6e9797ed-0ede-490a-ad13-a640600a5e9a","Type":"ContainerDied","Data":"c4019f5b4880889c474c9e2e7d457812e0296a8c8068daa31b5f96eef3185b3e"} Oct 08 00:59:48 crc kubenswrapper[4871]: I1008 00:59:48.127553 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:59:48 crc kubenswrapper[4871]: I1008 00:59:48.167146 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b944m/crc-debug-9cvsq"] Oct 08 00:59:48 crc kubenswrapper[4871]: I1008 00:59:48.180391 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b944m/crc-debug-9cvsq"] Oct 08 00:59:48 crc kubenswrapper[4871]: I1008 00:59:48.221420 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e9797ed-0ede-490a-ad13-a640600a5e9a-host\") pod \"6e9797ed-0ede-490a-ad13-a640600a5e9a\" (UID: \"6e9797ed-0ede-490a-ad13-a640600a5e9a\") " Oct 08 00:59:48 crc kubenswrapper[4871]: I1008 00:59:48.221487 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbt8h\" (UniqueName: \"kubernetes.io/projected/6e9797ed-0ede-490a-ad13-a640600a5e9a-kube-api-access-tbt8h\") pod \"6e9797ed-0ede-490a-ad13-a640600a5e9a\" (UID: \"6e9797ed-0ede-490a-ad13-a640600a5e9a\") " Oct 08 00:59:48 crc kubenswrapper[4871]: I1008 00:59:48.221553 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e9797ed-0ede-490a-ad13-a640600a5e9a-host" (OuterVolumeSpecName: "host") pod "6e9797ed-0ede-490a-ad13-a640600a5e9a" (UID: "6e9797ed-0ede-490a-ad13-a640600a5e9a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:59:48 crc kubenswrapper[4871]: I1008 00:59:48.222247 4871 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e9797ed-0ede-490a-ad13-a640600a5e9a-host\") on node \"crc\" DevicePath \"\"" Oct 08 00:59:48 crc kubenswrapper[4871]: I1008 00:59:48.228308 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e9797ed-0ede-490a-ad13-a640600a5e9a-kube-api-access-tbt8h" (OuterVolumeSpecName: "kube-api-access-tbt8h") pod "6e9797ed-0ede-490a-ad13-a640600a5e9a" (UID: "6e9797ed-0ede-490a-ad13-a640600a5e9a"). InnerVolumeSpecName "kube-api-access-tbt8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:59:48 crc kubenswrapper[4871]: I1008 00:59:48.324072 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbt8h\" (UniqueName: \"kubernetes.io/projected/6e9797ed-0ede-490a-ad13-a640600a5e9a-kube-api-access-tbt8h\") on node \"crc\" DevicePath \"\"" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.008460 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e9797ed-0ede-490a-ad13-a640600a5e9a" path="/var/lib/kubelet/pods/6e9797ed-0ede-490a-ad13-a640600a5e9a/volumes" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.014187 4871 scope.go:117] "RemoveContainer" containerID="c4019f5b4880889c474c9e2e7d457812e0296a8c8068daa31b5f96eef3185b3e" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.014287 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-9cvsq" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.372455 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b944m/crc-debug-khggr"] Oct 08 00:59:49 crc kubenswrapper[4871]: E1008 00:59:49.373460 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerName="extract-utilities" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.373484 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerName="extract-utilities" Oct 08 00:59:49 crc kubenswrapper[4871]: E1008 00:59:49.373513 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerName="registry-server" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.373526 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerName="registry-server" Oct 08 00:59:49 crc kubenswrapper[4871]: E1008 00:59:49.373562 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e9797ed-0ede-490a-ad13-a640600a5e9a" containerName="container-00" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.373576 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e9797ed-0ede-490a-ad13-a640600a5e9a" containerName="container-00" Oct 08 00:59:49 crc kubenswrapper[4871]: E1008 00:59:49.373607 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerName="registry-server" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.373620 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerName="registry-server" Oct 08 00:59:49 crc kubenswrapper[4871]: E1008 00:59:49.373667 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerName="extract-utilities" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.373682 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerName="extract-utilities" Oct 08 00:59:49 crc kubenswrapper[4871]: E1008 00:59:49.373702 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerName="extract-content" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.373715 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerName="extract-content" Oct 08 00:59:49 crc kubenswrapper[4871]: E1008 00:59:49.373740 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerName="extract-content" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.373753 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerName="extract-content" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.374148 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="539bc38f-ee93-4b12-a00c-ccd4ac10221d" containerName="registry-server" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.374195 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e9797ed-0ede-490a-ad13-a640600a5e9a" containerName="container-00" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.374231 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f8d6dfe-9200-4130-ba88-34934aa55ea0" containerName="registry-server" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.375552 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.377583 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-b944m"/"default-dockercfg-f47bl" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.454841 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-host\") pod \"crc-debug-khggr\" (UID: \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\") " pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.454923 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnd9k\" (UniqueName: \"kubernetes.io/projected/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-kube-api-access-vnd9k\") pod \"crc-debug-khggr\" (UID: \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\") " pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.556870 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-host\") pod \"crc-debug-khggr\" (UID: \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\") " pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.556962 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnd9k\" (UniqueName: \"kubernetes.io/projected/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-kube-api-access-vnd9k\") pod \"crc-debug-khggr\" (UID: \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\") " pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.557055 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-host\") pod \"crc-debug-khggr\" (UID: \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\") " pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.584821 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnd9k\" (UniqueName: \"kubernetes.io/projected/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-kube-api-access-vnd9k\") pod \"crc-debug-khggr\" (UID: \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\") " pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 00:59:49 crc kubenswrapper[4871]: I1008 00:59:49.694172 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 00:59:50 crc kubenswrapper[4871]: I1008 00:59:50.027353 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/crc-debug-khggr" event={"ID":"51e52abb-4ce4-4900-ac03-7be27b3b7fc2","Type":"ContainerStarted","Data":"06804e71303821f5dd85704929191cb0e1d15fd6478514154104cc7fcfbce9bc"} Oct 08 00:59:50 crc kubenswrapper[4871]: I1008 00:59:50.027714 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/crc-debug-khggr" event={"ID":"51e52abb-4ce4-4900-ac03-7be27b3b7fc2","Type":"ContainerStarted","Data":"2e4a1929bc6d10e23f872c0a7ea90459770a00cc6137fe29f01c49426ff33d74"} Oct 08 00:59:50 crc kubenswrapper[4871]: I1008 00:59:50.056853 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b944m/crc-debug-khggr" podStartSLOduration=1.056826155 podStartE2EDuration="1.056826155s" podCreationTimestamp="2025-10-08 00:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:59:50.043513015 +0000 UTC m=+10263.846211098" watchObservedRunningTime="2025-10-08 00:59:50.056826155 +0000 UTC m=+10263.859524258" Oct 08 00:59:51 crc kubenswrapper[4871]: I1008 00:59:51.040814 4871 generic.go:334] "Generic (PLEG): container finished" podID="51e52abb-4ce4-4900-ac03-7be27b3b7fc2" containerID="06804e71303821f5dd85704929191cb0e1d15fd6478514154104cc7fcfbce9bc" exitCode=0 Oct 08 00:59:51 crc kubenswrapper[4871]: I1008 00:59:51.040898 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/crc-debug-khggr" event={"ID":"51e52abb-4ce4-4900-ac03-7be27b3b7fc2","Type":"ContainerDied","Data":"06804e71303821f5dd85704929191cb0e1d15fd6478514154104cc7fcfbce9bc"} Oct 08 00:59:52 crc kubenswrapper[4871]: I1008 00:59:52.163652 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 00:59:52 crc kubenswrapper[4871]: I1008 00:59:52.308246 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-host\") pod \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\" (UID: \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\") " Oct 08 00:59:52 crc kubenswrapper[4871]: I1008 00:59:52.308362 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-host" (OuterVolumeSpecName: "host") pod "51e52abb-4ce4-4900-ac03-7be27b3b7fc2" (UID: "51e52abb-4ce4-4900-ac03-7be27b3b7fc2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:59:52 crc kubenswrapper[4871]: I1008 00:59:52.308415 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnd9k\" (UniqueName: \"kubernetes.io/projected/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-kube-api-access-vnd9k\") pod \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\" (UID: \"51e52abb-4ce4-4900-ac03-7be27b3b7fc2\") " Oct 08 00:59:52 crc kubenswrapper[4871]: I1008 00:59:52.309029 4871 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-host\") on node \"crc\" DevicePath \"\"" Oct 08 00:59:52 crc kubenswrapper[4871]: I1008 00:59:52.318112 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-kube-api-access-vnd9k" (OuterVolumeSpecName: "kube-api-access-vnd9k") pod "51e52abb-4ce4-4900-ac03-7be27b3b7fc2" (UID: "51e52abb-4ce4-4900-ac03-7be27b3b7fc2"). InnerVolumeSpecName "kube-api-access-vnd9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:59:52 crc kubenswrapper[4871]: I1008 00:59:52.410367 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnd9k\" (UniqueName: \"kubernetes.io/projected/51e52abb-4ce4-4900-ac03-7be27b3b7fc2-kube-api-access-vnd9k\") on node \"crc\" DevicePath \"\"" Oct 08 00:59:53 crc kubenswrapper[4871]: I1008 00:59:53.068554 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/crc-debug-khggr" event={"ID":"51e52abb-4ce4-4900-ac03-7be27b3b7fc2","Type":"ContainerDied","Data":"2e4a1929bc6d10e23f872c0a7ea90459770a00cc6137fe29f01c49426ff33d74"} Oct 08 00:59:53 crc kubenswrapper[4871]: I1008 00:59:53.068594 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e4a1929bc6d10e23f872c0a7ea90459770a00cc6137fe29f01c49426ff33d74" Oct 08 00:59:53 crc kubenswrapper[4871]: I1008 00:59:53.068663 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-khggr" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.164078 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc"] Oct 08 01:00:00 crc kubenswrapper[4871]: E1008 01:00:00.165064 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e52abb-4ce4-4900-ac03-7be27b3b7fc2" containerName="container-00" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.165077 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e52abb-4ce4-4900-ac03-7be27b3b7fc2" containerName="container-00" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.165312 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e52abb-4ce4-4900-ac03-7be27b3b7fc2" containerName="container-00" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.166197 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.168442 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.170241 4871 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.173326 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc"] Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.256183 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-config-volume\") pod \"collect-profiles-29331420-d2zgc\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.256216 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-secret-volume\") pod \"collect-profiles-29331420-d2zgc\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.256319 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7lb7\" (UniqueName: \"kubernetes.io/projected/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-kube-api-access-v7lb7\") pod \"collect-profiles-29331420-d2zgc\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.359218 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7lb7\" (UniqueName: \"kubernetes.io/projected/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-kube-api-access-v7lb7\") pod \"collect-profiles-29331420-d2zgc\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.359921 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-config-volume\") pod \"collect-profiles-29331420-d2zgc\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.359968 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-secret-volume\") pod \"collect-profiles-29331420-d2zgc\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.360950 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-config-volume\") pod \"collect-profiles-29331420-d2zgc\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.366646 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-secret-volume\") pod \"collect-profiles-29331420-d2zgc\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.375242 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7lb7\" (UniqueName: \"kubernetes.io/projected/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-kube-api-access-v7lb7\") pod \"collect-profiles-29331420-d2zgc\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.504297 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.654202 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b944m/crc-debug-khggr"] Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.662872 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b944m/crc-debug-khggr"] Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.975170 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc"] Oct 08 01:00:00 crc kubenswrapper[4871]: I1008 01:00:00.996528 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51e52abb-4ce4-4900-ac03-7be27b3b7fc2" path="/var/lib/kubelet/pods/51e52abb-4ce4-4900-ac03-7be27b3b7fc2/volumes" Oct 08 01:00:01 crc kubenswrapper[4871]: I1008 01:00:01.165222 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" event={"ID":"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e","Type":"ContainerStarted","Data":"31e36d290f1a20e156fef3d22139d7e7198e5161a1f9390651e0f7f78b743b13"} Oct 08 01:00:01 crc kubenswrapper[4871]: I1008 01:00:01.165548 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" event={"ID":"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e","Type":"ContainerStarted","Data":"2295187589a71c5e436bbf9b84a33c097809c81773ced91e630ddb3c7e67247a"} Oct 08 01:00:01 crc kubenswrapper[4871]: I1008 01:00:01.183317 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" podStartSLOduration=1.183298845 podStartE2EDuration="1.183298845s" podCreationTimestamp="2025-10-08 01:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 01:00:01.17852347 +0000 UTC m=+10274.981221543" watchObservedRunningTime="2025-10-08 01:00:01.183298845 +0000 UTC m=+10274.985996918" Oct 08 01:00:01 crc kubenswrapper[4871]: I1008 01:00:01.866116 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b944m/crc-debug-dwj56"] Oct 08 01:00:01 crc kubenswrapper[4871]: I1008 01:00:01.867661 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:01 crc kubenswrapper[4871]: I1008 01:00:01.870752 4871 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-b944m"/"default-dockercfg-f47bl" Oct 08 01:00:01 crc kubenswrapper[4871]: I1008 01:00:01.998095 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w575z\" (UniqueName: \"kubernetes.io/projected/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-kube-api-access-w575z\") pod \"crc-debug-dwj56\" (UID: \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\") " pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:01 crc kubenswrapper[4871]: I1008 01:00:01.998153 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-host\") pod \"crc-debug-dwj56\" (UID: \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\") " pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:02 crc kubenswrapper[4871]: I1008 01:00:02.100307 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w575z\" (UniqueName: \"kubernetes.io/projected/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-kube-api-access-w575z\") pod \"crc-debug-dwj56\" (UID: \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\") " pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:02 crc kubenswrapper[4871]: I1008 01:00:02.100613 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-host\") pod \"crc-debug-dwj56\" (UID: \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\") " pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:02 crc kubenswrapper[4871]: I1008 01:00:02.100772 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-host\") pod \"crc-debug-dwj56\" (UID: \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\") " pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:02 crc kubenswrapper[4871]: I1008 01:00:02.127650 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w575z\" (UniqueName: \"kubernetes.io/projected/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-kube-api-access-w575z\") pod \"crc-debug-dwj56\" (UID: \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\") " pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:02 crc kubenswrapper[4871]: I1008 01:00:02.176201 4871 generic.go:334] "Generic (PLEG): container finished" podID="be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e" containerID="31e36d290f1a20e156fef3d22139d7e7198e5161a1f9390651e0f7f78b743b13" exitCode=0 Oct 08 01:00:02 crc kubenswrapper[4871]: I1008 01:00:02.176239 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" event={"ID":"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e","Type":"ContainerDied","Data":"31e36d290f1a20e156fef3d22139d7e7198e5161a1f9390651e0f7f78b743b13"} Oct 08 01:00:02 crc kubenswrapper[4871]: I1008 01:00:02.189610 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:02 crc kubenswrapper[4871]: W1008 01:00:02.236065 4871 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b209bbb_488d_4fea_bbb6_ae3899aad3ee.slice/crio-3fbceedaabcaa31f16197d0b1a2e003d4bb17d7ce7edf1de2e29385cbfd43b9f WatchSource:0}: Error finding container 3fbceedaabcaa31f16197d0b1a2e003d4bb17d7ce7edf1de2e29385cbfd43b9f: Status 404 returned error can't find the container with id 3fbceedaabcaa31f16197d0b1a2e003d4bb17d7ce7edf1de2e29385cbfd43b9f Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.188303 4871 generic.go:334] "Generic (PLEG): container finished" podID="2b209bbb-488d-4fea-bbb6-ae3899aad3ee" containerID="f297de073eecadddbe59f84fcdd93711469d876767af77d1da156e3285192361" exitCode=0 Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.188436 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/crc-debug-dwj56" event={"ID":"2b209bbb-488d-4fea-bbb6-ae3899aad3ee","Type":"ContainerDied","Data":"f297de073eecadddbe59f84fcdd93711469d876767af77d1da156e3285192361"} Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.188927 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/crc-debug-dwj56" event={"ID":"2b209bbb-488d-4fea-bbb6-ae3899aad3ee","Type":"ContainerStarted","Data":"3fbceedaabcaa31f16197d0b1a2e003d4bb17d7ce7edf1de2e29385cbfd43b9f"} Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.243491 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b944m/crc-debug-dwj56"] Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.256478 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b944m/crc-debug-dwj56"] Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.632799 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.740120 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-config-volume\") pod \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.740270 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-secret-volume\") pod \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.740309 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7lb7\" (UniqueName: \"kubernetes.io/projected/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-kube-api-access-v7lb7\") pod \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\" (UID: \"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e\") " Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.740812 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-config-volume" (OuterVolumeSpecName: "config-volume") pod "be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e" (UID: "be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.751025 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-kube-api-access-v7lb7" (OuterVolumeSpecName: "kube-api-access-v7lb7") pod "be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e" (UID: "be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e"). InnerVolumeSpecName "kube-api-access-v7lb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.753144 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e" (UID: "be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.843011 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7lb7\" (UniqueName: \"kubernetes.io/projected/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-kube-api-access-v7lb7\") on node \"crc\" DevicePath \"\"" Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.843064 4871 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 01:00:03 crc kubenswrapper[4871]: I1008 01:00:03.843079 4871 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.216518 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.216524 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331420-d2zgc" event={"ID":"be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e","Type":"ContainerDied","Data":"2295187589a71c5e436bbf9b84a33c097809c81773ced91e630ddb3c7e67247a"} Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.216698 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2295187589a71c5e436bbf9b84a33c097809c81773ced91e630ddb3c7e67247a" Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.277229 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg"] Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.288603 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331375-5gmrg"] Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.309361 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.354373 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w575z\" (UniqueName: \"kubernetes.io/projected/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-kube-api-access-w575z\") pod \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\" (UID: \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\") " Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.354521 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-host\") pod \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\" (UID: \"2b209bbb-488d-4fea-bbb6-ae3899aad3ee\") " Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.355234 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-host" (OuterVolumeSpecName: "host") pod "2b209bbb-488d-4fea-bbb6-ae3899aad3ee" (UID: "2b209bbb-488d-4fea-bbb6-ae3899aad3ee"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.355779 4871 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-host\") on node \"crc\" DevicePath \"\"" Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.362633 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-kube-api-access-w575z" (OuterVolumeSpecName: "kube-api-access-w575z") pod "2b209bbb-488d-4fea-bbb6-ae3899aad3ee" (UID: "2b209bbb-488d-4fea-bbb6-ae3899aad3ee"). InnerVolumeSpecName "kube-api-access-w575z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 01:00:04 crc kubenswrapper[4871]: I1008 01:00:04.457906 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w575z\" (UniqueName: \"kubernetes.io/projected/2b209bbb-488d-4fea-bbb6-ae3899aad3ee-kube-api-access-w575z\") on node \"crc\" DevicePath \"\"" Oct 08 01:00:05 crc kubenswrapper[4871]: I1008 01:00:05.007290 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0866e481-6790-4024-b638-0e86221234eb" path="/var/lib/kubelet/pods/0866e481-6790-4024-b638-0e86221234eb/volumes" Oct 08 01:00:05 crc kubenswrapper[4871]: I1008 01:00:05.009373 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b209bbb-488d-4fea-bbb6-ae3899aad3ee" path="/var/lib/kubelet/pods/2b209bbb-488d-4fea-bbb6-ae3899aad3ee/volumes" Oct 08 01:00:05 crc kubenswrapper[4871]: I1008 01:00:05.232771 4871 scope.go:117] "RemoveContainer" containerID="f297de073eecadddbe59f84fcdd93711469d876767af77d1da156e3285192361" Oct 08 01:00:05 crc kubenswrapper[4871]: I1008 01:00:05.232851 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/crc-debug-dwj56" Oct 08 01:00:35 crc kubenswrapper[4871]: I1008 01:00:35.511917 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 01:00:35 crc kubenswrapper[4871]: I1008 01:00:35.512491 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 01:00:57 crc kubenswrapper[4871]: I1008 01:00:57.382800 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp_fe082719-5871-4673-b838-cf19f5b6cd45/util/0.log" Oct 08 01:00:57 crc kubenswrapper[4871]: I1008 01:00:57.529035 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp_fe082719-5871-4673-b838-cf19f5b6cd45/util/0.log" Oct 08 01:00:57 crc kubenswrapper[4871]: I1008 01:00:57.553047 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp_fe082719-5871-4673-b838-cf19f5b6cd45/pull/0.log" Oct 08 01:00:57 crc kubenswrapper[4871]: I1008 01:00:57.564015 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp_fe082719-5871-4673-b838-cf19f5b6cd45/pull/0.log" Oct 08 01:00:57 crc kubenswrapper[4871]: I1008 01:00:57.756023 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp_fe082719-5871-4673-b838-cf19f5b6cd45/pull/0.log" Oct 08 01:00:57 crc kubenswrapper[4871]: I1008 01:00:57.762324 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp_fe082719-5871-4673-b838-cf19f5b6cd45/extract/0.log" Oct 08 01:00:57 crc kubenswrapper[4871]: I1008 01:00:57.771172 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e60777004095lnxxp_fe082719-5871-4673-b838-cf19f5b6cd45/util/0.log" Oct 08 01:00:57 crc kubenswrapper[4871]: I1008 01:00:57.919249 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f56ff694-fh5vz_07b6d737-7f0c-494d-adfe-0a641261c344/kube-rbac-proxy/0.log" Oct 08 01:00:57 crc kubenswrapper[4871]: I1008 01:00:57.990514 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-v6x97_b58f786d-7c8a-4506-94d4-bf1fcf543871/kube-rbac-proxy/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.038446 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f56ff694-fh5vz_07b6d737-7f0c-494d-adfe-0a641261c344/manager/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.157694 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-v6x97_b58f786d-7c8a-4506-94d4-bf1fcf543871/manager/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.184898 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-5ndrn_e65e1c1c-1bbc-488e-821c-790bb85bcd55/kube-rbac-proxy/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.215924 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-5ndrn_e65e1c1c-1bbc-488e-821c-790bb85bcd55/manager/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.329871 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-fd648f65-fsrt7_e85a9c1c-2748-4ae7-a4fa-cfb555348feb/kube-rbac-proxy/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.497514 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-7ccfc8cf49-lg6tk_8acf9fc2-6307-4d04-921c-cdfc851f8c52/kube-rbac-proxy/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.513907 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-fd648f65-fsrt7_e85a9c1c-2748-4ae7-a4fa-cfb555348feb/manager/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.550614 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-7ccfc8cf49-lg6tk_8acf9fc2-6307-4d04-921c-cdfc851f8c52/manager/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.735761 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b477879bc-jd9ql_6916e675-a275-4e0f-b43d-c40a8973f13f/manager/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.763389 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b477879bc-jd9ql_6916e675-a275-4e0f-b43d-c40a8973f13f/kube-rbac-proxy/0.log" Oct 08 01:00:58 crc kubenswrapper[4871]: I1008 01:00:58.808078 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-c6bxk_e08fd5ec-5b08-4ad9-b298-886bda95d3d9/kube-rbac-proxy/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.065855 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5467f8988c-ftlvg_63b31dae-e86d-4735-96b8-125a9a066a3e/kube-rbac-proxy/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.100885 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5467f8988c-ftlvg_63b31dae-e86d-4735-96b8-125a9a066a3e/manager/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.173555 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-c6bxk_e08fd5ec-5b08-4ad9-b298-886bda95d3d9/manager/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.283173 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b84cc7657-fwjnb_b3a4e444-6508-4ae3-8b40-3f238f252ed2/kube-rbac-proxy/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.397157 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-rtfnp_0c7c4f04-05a9-40a5-9cb1-9320e33444b1/kube-rbac-proxy/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.413339 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b84cc7657-fwjnb_b3a4e444-6508-4ae3-8b40-3f238f252ed2/manager/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.489844 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-rtfnp_0c7c4f04-05a9-40a5-9cb1-9320e33444b1/manager/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.586628 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-8bw69_cda39518-2b25-43e2-8afa-b0ff0ec18ff7/kube-rbac-proxy/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.657916 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-8bw69_cda39518-2b25-43e2-8afa-b0ff0ec18ff7/manager/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.764228 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-czwls_fc43555d-cedf-469e-af48-b2f3cce766c0/kube-rbac-proxy/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.806549 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-czwls_fc43555d-cedf-469e-af48-b2f3cce766c0/manager/0.log" Oct 08 01:00:59 crc kubenswrapper[4871]: I1008 01:00:59.907329 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-hk9db_36980ce9-53e4-401a-b093-bd1ba933ae83/kube-rbac-proxy/0.log" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.031207 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-qj54s_9772f309-5c98-45d7-a404-b14c2ab199e6/kube-rbac-proxy/0.log" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.056863 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-hk9db_36980ce9-53e4-401a-b093-bd1ba933ae83/manager/0.log" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.159519 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29331421-75nf7"] Oct 08 01:01:00 crc kubenswrapper[4871]: E1008 01:01:00.173853 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b209bbb-488d-4fea-bbb6-ae3899aad3ee" containerName="container-00" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.173892 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b209bbb-488d-4fea-bbb6-ae3899aad3ee" containerName="container-00" Oct 08 01:01:00 crc kubenswrapper[4871]: E1008 01:01:00.173913 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e" containerName="collect-profiles" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.173921 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e" containerName="collect-profiles" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.174206 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b209bbb-488d-4fea-bbb6-ae3899aad3ee" containerName="container-00" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.174226 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="be2aaa01-8e3e-4cb4-ac7e-633e8b4f505e" containerName="collect-profiles" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.176868 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.197993 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29331421-75nf7"] Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.229486 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-qj54s_9772f309-5c98-45d7-a404-b14c2ab199e6/manager/0.log" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.271178 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txtrb\" (UniqueName: \"kubernetes.io/projected/309d839b-d676-47dd-adf8-affe46e4c807-kube-api-access-txtrb\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.271214 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-combined-ca-bundle\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.271243 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-fernet-keys\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.271323 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-config-data\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.274605 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6875c66686wj76g_f919f707-59de-4f00-a9cf-177ed5696694/kube-rbac-proxy/0.log" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.335965 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6875c66686wj76g_f919f707-59de-4f00-a9cf-177ed5696694/manager/0.log" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.372550 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txtrb\" (UniqueName: \"kubernetes.io/projected/309d839b-d676-47dd-adf8-affe46e4c807-kube-api-access-txtrb\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.372591 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-combined-ca-bundle\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.372617 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-fernet-keys\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.372703 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-config-data\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.380180 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-config-data\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.386921 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-fernet-keys\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.393528 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txtrb\" (UniqueName: \"kubernetes.io/projected/309d839b-d676-47dd-adf8-affe46e4c807-kube-api-access-txtrb\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.397446 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-combined-ca-bundle\") pod \"keystone-cron-29331421-75nf7\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.425066 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6bfd56c677-s7qzs_15bcef66-2998-4b89-b4ad-18e06aaf57b6/kube-rbac-proxy/0.log" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.503509 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.700904 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-b6d857f89-scc7s_6198e2f8-88e4-4ad8-bda0-aab47a10f07a/kube-rbac-proxy/0.log" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.810221 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-b6d857f89-scc7s_6198e2f8-88e4-4ad8-bda0-aab47a10f07a/operator/0.log" Oct 08 01:01:00 crc kubenswrapper[4871]: I1008 01:01:00.995866 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29331421-75nf7"] Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.028231 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-cgtsg_e8cf523f-4104-44ad-b446-86ad5318cc9f/registry-server/0.log" Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.121837 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54d485fd9-c5989_f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6/kube-rbac-proxy/0.log" Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.237513 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54d485fd9-c5989_f4c2f2ed-78b8-40be-9d35-6f1cb7f719c6/manager/0.log" Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.334668 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-ht58x_0521e121-6b1f-44e3-9823-73f5cb04b5dd/kube-rbac-proxy/0.log" Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.425757 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-ht58x_0521e121-6b1f-44e3-9823-73f5cb04b5dd/manager/0.log" Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.506815 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-q7hjx_efa19466-3280-43e2-b4a1-2d2077114241/operator/0.log" Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.611684 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-nz284_e548d82a-0d88-4597-a73e-5e6230e58081/kube-rbac-proxy/0.log" Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.733177 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-nz284_e548d82a-0d88-4597-a73e-5e6230e58081/manager/0.log" Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.908330 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-hvgqm_6a176a41-220a-4472-9ca5-4e341f86ce49/kube-rbac-proxy/0.log" Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.958875 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331421-75nf7" event={"ID":"309d839b-d676-47dd-adf8-affe46e4c807","Type":"ContainerStarted","Data":"0262f80cc489ede61158012369df4e0d7327ca2528c93feae355740b92f89183"} Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.958915 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331421-75nf7" event={"ID":"309d839b-d676-47dd-adf8-affe46e4c807","Type":"ContainerStarted","Data":"821e8b87abccdc77d5809a1398ae52b7eb9e1e6eb19162666499b18ab094106e"} Oct 08 01:01:01 crc kubenswrapper[4871]: I1008 01:01:01.981875 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29331421-75nf7" podStartSLOduration=1.981856803 podStartE2EDuration="1.981856803s" podCreationTimestamp="2025-10-08 01:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 01:01:01.97147993 +0000 UTC m=+10335.774178003" watchObservedRunningTime="2025-10-08 01:01:01.981856803 +0000 UTC m=+10335.784554876" Oct 08 01:01:02 crc kubenswrapper[4871]: I1008 01:01:02.014354 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-n5vms_aab600a0-d647-44e5-95a4-870d422385ca/kube-rbac-proxy/0.log" Oct 08 01:01:02 crc kubenswrapper[4871]: I1008 01:01:02.109362 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-n5vms_aab600a0-d647-44e5-95a4-870d422385ca/manager/0.log" Oct 08 01:01:02 crc kubenswrapper[4871]: I1008 01:01:02.222213 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-h5h7d_ad701ba2-f685-4f23-a42a-dc2480b2e92b/kube-rbac-proxy/0.log" Oct 08 01:01:02 crc kubenswrapper[4871]: I1008 01:01:02.234723 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-hvgqm_6a176a41-220a-4472-9ca5-4e341f86ce49/manager/0.log" Oct 08 01:01:02 crc kubenswrapper[4871]: I1008 01:01:02.314808 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-h5h7d_ad701ba2-f685-4f23-a42a-dc2480b2e92b/manager/0.log" Oct 08 01:01:02 crc kubenswrapper[4871]: I1008 01:01:02.831712 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6bfd56c677-s7qzs_15bcef66-2998-4b89-b4ad-18e06aaf57b6/manager/0.log" Oct 08 01:01:03 crc kubenswrapper[4871]: I1008 01:01:03.981499 4871 generic.go:334] "Generic (PLEG): container finished" podID="309d839b-d676-47dd-adf8-affe46e4c807" containerID="0262f80cc489ede61158012369df4e0d7327ca2528c93feae355740b92f89183" exitCode=0 Oct 08 01:01:03 crc kubenswrapper[4871]: I1008 01:01:03.981557 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331421-75nf7" event={"ID":"309d839b-d676-47dd-adf8-affe46e4c807","Type":"ContainerDied","Data":"0262f80cc489ede61158012369df4e0d7327ca2528c93feae355740b92f89183"} Oct 08 01:01:04 crc kubenswrapper[4871]: I1008 01:01:04.756687 4871 scope.go:117] "RemoveContainer" containerID="9dae4e48919efc401d480c7a543d30e5fcf3a5dbc75164eea3d458358fc6a280" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.416011 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.476732 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-fernet-keys\") pod \"309d839b-d676-47dd-adf8-affe46e4c807\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.476985 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txtrb\" (UniqueName: \"kubernetes.io/projected/309d839b-d676-47dd-adf8-affe46e4c807-kube-api-access-txtrb\") pod \"309d839b-d676-47dd-adf8-affe46e4c807\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.477011 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-combined-ca-bundle\") pod \"309d839b-d676-47dd-adf8-affe46e4c807\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.477071 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-config-data\") pod \"309d839b-d676-47dd-adf8-affe46e4c807\" (UID: \"309d839b-d676-47dd-adf8-affe46e4c807\") " Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.482229 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "309d839b-d676-47dd-adf8-affe46e4c807" (UID: "309d839b-d676-47dd-adf8-affe46e4c807"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.482520 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/309d839b-d676-47dd-adf8-affe46e4c807-kube-api-access-txtrb" (OuterVolumeSpecName: "kube-api-access-txtrb") pod "309d839b-d676-47dd-adf8-affe46e4c807" (UID: "309d839b-d676-47dd-adf8-affe46e4c807"). InnerVolumeSpecName "kube-api-access-txtrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.511960 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "309d839b-d676-47dd-adf8-affe46e4c807" (UID: "309d839b-d676-47dd-adf8-affe46e4c807"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.512239 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.512276 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.540121 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-config-data" (OuterVolumeSpecName: "config-data") pod "309d839b-d676-47dd-adf8-affe46e4c807" (UID: "309d839b-d676-47dd-adf8-affe46e4c807"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.579121 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txtrb\" (UniqueName: \"kubernetes.io/projected/309d839b-d676-47dd-adf8-affe46e4c807-kube-api-access-txtrb\") on node \"crc\" DevicePath \"\"" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.579151 4871 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.579161 4871 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 01:01:05 crc kubenswrapper[4871]: I1008 01:01:05.579169 4871 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/309d839b-d676-47dd-adf8-affe46e4c807-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 01:01:06 crc kubenswrapper[4871]: I1008 01:01:06.004712 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331421-75nf7" event={"ID":"309d839b-d676-47dd-adf8-affe46e4c807","Type":"ContainerDied","Data":"821e8b87abccdc77d5809a1398ae52b7eb9e1e6eb19162666499b18ab094106e"} Oct 08 01:01:06 crc kubenswrapper[4871]: I1008 01:01:06.004951 4871 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="821e8b87abccdc77d5809a1398ae52b7eb9e1e6eb19162666499b18ab094106e" Oct 08 01:01:06 crc kubenswrapper[4871]: I1008 01:01:06.004762 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331421-75nf7" Oct 08 01:01:19 crc kubenswrapper[4871]: I1008 01:01:19.881592 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-cc8tr_18a8f350-7724-4302-8663-499480b9af6e/control-plane-machine-set-operator/0.log" Oct 08 01:01:20 crc kubenswrapper[4871]: I1008 01:01:20.084901 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2zz24_a6b7f864-334b-4d77-a8e4-bf62b614fb80/kube-rbac-proxy/0.log" Oct 08 01:01:20 crc kubenswrapper[4871]: I1008 01:01:20.171052 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2zz24_a6b7f864-334b-4d77-a8e4-bf62b614fb80/machine-api-operator/0.log" Oct 08 01:01:33 crc kubenswrapper[4871]: I1008 01:01:33.850587 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-p5q48_4a764d87-e672-42a7-b88c-d1485bfd7c19/cert-manager-controller/0.log" Oct 08 01:01:34 crc kubenswrapper[4871]: I1008 01:01:34.004667 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-9x9xp_7897ec6d-8c3f-4f9b-93b2-489e21dc01af/cert-manager-cainjector/0.log" Oct 08 01:01:34 crc kubenswrapper[4871]: I1008 01:01:34.028842 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-zls94_90172f12-e077-4e3a-93a0-115a93588eac/cert-manager-webhook/0.log" Oct 08 01:01:35 crc kubenswrapper[4871]: I1008 01:01:35.512900 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 01:01:35 crc kubenswrapper[4871]: I1008 01:01:35.514103 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 01:01:35 crc kubenswrapper[4871]: I1008 01:01:35.514248 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 01:01:35 crc kubenswrapper[4871]: I1008 01:01:35.515272 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6b5bdcf7675888de17bfd75524c8b30d3e8713d1d2a3c5cd0e1017c4ecc59890"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 01:01:35 crc kubenswrapper[4871]: I1008 01:01:35.515441 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://6b5bdcf7675888de17bfd75524c8b30d3e8713d1d2a3c5cd0e1017c4ecc59890" gracePeriod=600 Oct 08 01:01:36 crc kubenswrapper[4871]: I1008 01:01:36.361942 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="6b5bdcf7675888de17bfd75524c8b30d3e8713d1d2a3c5cd0e1017c4ecc59890" exitCode=0 Oct 08 01:01:36 crc kubenswrapper[4871]: I1008 01:01:36.361978 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"6b5bdcf7675888de17bfd75524c8b30d3e8713d1d2a3c5cd0e1017c4ecc59890"} Oct 08 01:01:36 crc kubenswrapper[4871]: I1008 01:01:36.362964 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerStarted","Data":"07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523"} Oct 08 01:01:36 crc kubenswrapper[4871]: I1008 01:01:36.363007 4871 scope.go:117] "RemoveContainer" containerID="58b82ad792dd8ab19a18108f2e00d76f269bb1257655ad3634141ea79f71a807" Oct 08 01:01:47 crc kubenswrapper[4871]: I1008 01:01:47.186198 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-vs759_fe9e41ea-e4e6-4c17-bd98-50ad8216b160/nmstate-console-plugin/0.log" Oct 08 01:01:47 crc kubenswrapper[4871]: I1008 01:01:47.403400 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gvgbf_b22d7c95-54d4-4fd5-87fa-3e6cf69a6932/nmstate-handler/0.log" Oct 08 01:01:47 crc kubenswrapper[4871]: I1008 01:01:47.430574 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-hfbhn_058ddfce-22d3-45ea-8f8a-9463858b9d64/kube-rbac-proxy/0.log" Oct 08 01:01:47 crc kubenswrapper[4871]: I1008 01:01:47.451151 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-hfbhn_058ddfce-22d3-45ea-8f8a-9463858b9d64/nmstate-metrics/0.log" Oct 08 01:01:47 crc kubenswrapper[4871]: I1008 01:01:47.577104 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-lvfxx_8c4cee14-12e0-4ce6-9141-cfa251152a35/nmstate-operator/0.log" Oct 08 01:01:47 crc kubenswrapper[4871]: I1008 01:01:47.642029 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-v8gn8_11e73801-01a4-44c0-aaf8-5ad1e2af2239/nmstate-webhook/0.log" Oct 08 01:02:02 crc kubenswrapper[4871]: I1008 01:02:02.763936 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-gm89v_00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578/kube-rbac-proxy/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.003221 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-frr-files/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.184226 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-reloader/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.197032 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-frr-files/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.236412 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-gm89v_00fbeb58-1ed2-4c11-9cdf-48b9c4e7e578/controller/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.263700 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-metrics/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.356333 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-reloader/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.517336 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-metrics/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.518876 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-frr-files/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.523084 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-metrics/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.563770 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-reloader/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.687693 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-frr-files/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.706349 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-metrics/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.744314 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/cp-reloader/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.763351 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/controller/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.921543 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/kube-rbac-proxy/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.926701 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/frr-metrics/0.log" Oct 08 01:02:03 crc kubenswrapper[4871]: I1008 01:02:03.943000 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/kube-rbac-proxy-frr/0.log" Oct 08 01:02:04 crc kubenswrapper[4871]: I1008 01:02:04.110729 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/reloader/0.log" Oct 08 01:02:04 crc kubenswrapper[4871]: I1008 01:02:04.135369 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-2fc4m_f2649d22-c9ed-4d54-b85f-7739a2b50db2/frr-k8s-webhook-server/0.log" Oct 08 01:02:04 crc kubenswrapper[4871]: I1008 01:02:04.468984 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-b78c56f4c-qs2vb_840548f4-0e99-418d-98d4-400e684775e1/manager/0.log" Oct 08 01:02:04 crc kubenswrapper[4871]: I1008 01:02:04.619014 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-76fdd8f567-7nxxv_ad41768e-7589-4b4e-8193-2b48da644e53/webhook-server/0.log" Oct 08 01:02:04 crc kubenswrapper[4871]: I1008 01:02:04.713354 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4hjwr_f494e98b-a5ac-4f64-8f30-4da19c9bc9f7/kube-rbac-proxy/0.log" Oct 08 01:02:05 crc kubenswrapper[4871]: I1008 01:02:05.567149 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4hjwr_f494e98b-a5ac-4f64-8f30-4da19c9bc9f7/speaker/0.log" Oct 08 01:02:06 crc kubenswrapper[4871]: I1008 01:02:06.985603 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ksdzm_7c3493d5-0f90-4c6b-8ebc-ab2a78082037/frr/0.log" Oct 08 01:02:19 crc kubenswrapper[4871]: I1008 01:02:19.771312 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt_fe65ed5d-7ec6-47e4-a99a-9c214d055a33/util/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.007740 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt_fe65ed5d-7ec6-47e4-a99a-9c214d055a33/pull/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.008545 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt_fe65ed5d-7ec6-47e4-a99a-9c214d055a33/pull/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.033246 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt_fe65ed5d-7ec6-47e4-a99a-9c214d055a33/util/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.211978 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt_fe65ed5d-7ec6-47e4-a99a-9c214d055a33/util/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.249435 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt_fe65ed5d-7ec6-47e4-a99a-9c214d055a33/extract/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.258254 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69vxdrt_fe65ed5d-7ec6-47e4-a99a-9c214d055a33/pull/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.630066 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww_f0bdd976-9611-4335-93b3-6dec20c06076/util/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.878840 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww_f0bdd976-9611-4335-93b3-6dec20c06076/pull/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.878947 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww_f0bdd976-9611-4335-93b3-6dec20c06076/pull/0.log" Oct 08 01:02:20 crc kubenswrapper[4871]: I1008 01:02:20.957639 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww_f0bdd976-9611-4335-93b3-6dec20c06076/util/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.058522 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww_f0bdd976-9611-4335-93b3-6dec20c06076/util/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.107506 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww_f0bdd976-9611-4335-93b3-6dec20c06076/pull/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.157696 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cwjww_f0bdd976-9611-4335-93b3-6dec20c06076/extract/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.278825 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn_0ee6dc03-a9cc-468f-aea1-0ec712237561/util/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.428728 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn_0ee6dc03-a9cc-468f-aea1-0ec712237561/util/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.440618 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn_0ee6dc03-a9cc-468f-aea1-0ec712237561/pull/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.454214 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn_0ee6dc03-a9cc-468f-aea1-0ec712237561/pull/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.597804 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn_0ee6dc03-a9cc-468f-aea1-0ec712237561/pull/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.654945 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn_0ee6dc03-a9cc-468f-aea1-0ec712237561/extract/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.669338 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwc7wn_0ee6dc03-a9cc-468f-aea1-0ec712237561/util/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.791556 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-brt8h_bc1e7f6c-0af7-462e-8972-3251f7b091e7/extract-utilities/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.969618 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-brt8h_bc1e7f6c-0af7-462e-8972-3251f7b091e7/extract-content/0.log" Oct 08 01:02:21 crc kubenswrapper[4871]: I1008 01:02:21.969731 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-brt8h_bc1e7f6c-0af7-462e-8972-3251f7b091e7/extract-content/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.004653 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-brt8h_bc1e7f6c-0af7-462e-8972-3251f7b091e7/extract-utilities/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.122893 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-brt8h_bc1e7f6c-0af7-462e-8972-3251f7b091e7/extract-utilities/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.151055 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-brt8h_bc1e7f6c-0af7-462e-8972-3251f7b091e7/extract-content/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.292602 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-brt8h_bc1e7f6c-0af7-462e-8972-3251f7b091e7/registry-server/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.302539 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn6qm_e67ed465-27ab-4f9b-a8ab-faf4b17f7437/extract-utilities/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.468720 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn6qm_e67ed465-27ab-4f9b-a8ab-faf4b17f7437/extract-content/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.493355 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn6qm_e67ed465-27ab-4f9b-a8ab-faf4b17f7437/extract-utilities/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.520285 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn6qm_e67ed465-27ab-4f9b-a8ab-faf4b17f7437/extract-content/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.667224 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn6qm_e67ed465-27ab-4f9b-a8ab-faf4b17f7437/extract-utilities/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.691273 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn6qm_e67ed465-27ab-4f9b-a8ab-faf4b17f7437/extract-content/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.751142 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m_a34d9156-1901-41f3-a16c-c80d992569bb/util/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.916060 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m_a34d9156-1901-41f3-a16c-c80d992569bb/pull/0.log" Oct 08 01:02:22 crc kubenswrapper[4871]: I1008 01:02:22.976424 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m_a34d9156-1901-41f3-a16c-c80d992569bb/util/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.050293 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m_a34d9156-1901-41f3-a16c-c80d992569bb/pull/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.204300 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m_a34d9156-1901-41f3-a16c-c80d992569bb/pull/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.253671 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m_a34d9156-1901-41f3-a16c-c80d992569bb/util/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.258997 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckhc6m_a34d9156-1901-41f3-a16c-c80d992569bb/extract/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.454529 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kgzf8_464efd4a-2401-457e-bf72-41ce7b9f2cac/marketplace-operator/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.507366 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lggxl_8f7a029e-6cb3-4051-b2b2-a393044a2ea3/extract-utilities/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.679590 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lggxl_8f7a029e-6cb3-4051-b2b2-a393044a2ea3/extract-utilities/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.692728 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lggxl_8f7a029e-6cb3-4051-b2b2-a393044a2ea3/extract-content/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.784290 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lggxl_8f7a029e-6cb3-4051-b2b2-a393044a2ea3/extract-content/0.log" Oct 08 01:02:23 crc kubenswrapper[4871]: I1008 01:02:23.973114 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lggxl_8f7a029e-6cb3-4051-b2b2-a393044a2ea3/extract-utilities/0.log" Oct 08 01:02:24 crc kubenswrapper[4871]: I1008 01:02:24.002183 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lggxl_8f7a029e-6cb3-4051-b2b2-a393044a2ea3/extract-content/0.log" Oct 08 01:02:24 crc kubenswrapper[4871]: I1008 01:02:24.214244 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tsvn2_cef3b48f-0bfa-43f7-80ce-e58b5353637a/extract-utilities/0.log" Oct 08 01:02:24 crc kubenswrapper[4871]: I1008 01:02:24.224897 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn6qm_e67ed465-27ab-4f9b-a8ab-faf4b17f7437/registry-server/0.log" Oct 08 01:02:24 crc kubenswrapper[4871]: I1008 01:02:24.305776 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lggxl_8f7a029e-6cb3-4051-b2b2-a393044a2ea3/registry-server/0.log" Oct 08 01:02:24 crc kubenswrapper[4871]: I1008 01:02:24.412294 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tsvn2_cef3b48f-0bfa-43f7-80ce-e58b5353637a/extract-utilities/0.log" Oct 08 01:02:24 crc kubenswrapper[4871]: I1008 01:02:24.413269 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tsvn2_cef3b48f-0bfa-43f7-80ce-e58b5353637a/extract-content/0.log" Oct 08 01:02:24 crc kubenswrapper[4871]: I1008 01:02:24.416366 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tsvn2_cef3b48f-0bfa-43f7-80ce-e58b5353637a/extract-content/0.log" Oct 08 01:02:24 crc kubenswrapper[4871]: I1008 01:02:24.580406 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tsvn2_cef3b48f-0bfa-43f7-80ce-e58b5353637a/extract-utilities/0.log" Oct 08 01:02:24 crc kubenswrapper[4871]: I1008 01:02:24.620291 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tsvn2_cef3b48f-0bfa-43f7-80ce-e58b5353637a/extract-content/0.log" Oct 08 01:02:25 crc kubenswrapper[4871]: I1008 01:02:25.693921 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tsvn2_cef3b48f-0bfa-43f7-80ce-e58b5353637a/registry-server/0.log" Oct 08 01:02:38 crc kubenswrapper[4871]: I1008 01:02:38.465795 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-nr674_a3adbb28-7bd8-418c-94e4-e0a1ff7c8726/prometheus-operator/0.log" Oct 08 01:02:38 crc kubenswrapper[4871]: I1008 01:02:38.623516 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-8497998f8-ggr6m_1d3e2b5e-61e3-45ec-b39b-ec6aee52c207/prometheus-operator-admission-webhook/0.log" Oct 08 01:02:38 crc kubenswrapper[4871]: I1008 01:02:38.686119 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-8497998f8-hrq9c_caf90bb9-05c5-40da-be2d-b1c57ef8075f/prometheus-operator-admission-webhook/0.log" Oct 08 01:02:38 crc kubenswrapper[4871]: I1008 01:02:38.776265 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-ckv6s_07703b22-45ab-4158-9afe-a11f5f0f1e41/operator/0.log" Oct 08 01:02:38 crc kubenswrapper[4871]: I1008 01:02:38.877505 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-cjmdr_48ff787b-84fa-4f90-9d13-e853c115d1c8/perses-operator/0.log" Oct 08 01:02:53 crc kubenswrapper[4871]: E1008 01:02:53.411031 4871 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.68:34020->38.102.83.68:41109: write tcp 38.102.83.68:34020->38.102.83.68:41109: write: broken pipe Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.173477 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sljlg"] Oct 08 01:02:59 crc kubenswrapper[4871]: E1008 01:02:59.174536 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309d839b-d676-47dd-adf8-affe46e4c807" containerName="keystone-cron" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.174555 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="309d839b-d676-47dd-adf8-affe46e4c807" containerName="keystone-cron" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.174844 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="309d839b-d676-47dd-adf8-affe46e4c807" containerName="keystone-cron" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.176303 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.219850 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sljlg"] Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.302681 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gclm5\" (UniqueName: \"kubernetes.io/projected/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-kube-api-access-gclm5\") pod \"redhat-operators-sljlg\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.302744 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-utilities\") pod \"redhat-operators-sljlg\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.302908 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-catalog-content\") pod \"redhat-operators-sljlg\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.404446 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-catalog-content\") pod \"redhat-operators-sljlg\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.404520 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gclm5\" (UniqueName: \"kubernetes.io/projected/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-kube-api-access-gclm5\") pod \"redhat-operators-sljlg\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.404561 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-utilities\") pod \"redhat-operators-sljlg\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.405004 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-utilities\") pod \"redhat-operators-sljlg\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.405211 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-catalog-content\") pod \"redhat-operators-sljlg\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.424379 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gclm5\" (UniqueName: \"kubernetes.io/projected/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-kube-api-access-gclm5\") pod \"redhat-operators-sljlg\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:02:59 crc kubenswrapper[4871]: I1008 01:02:59.496732 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:03:00 crc kubenswrapper[4871]: I1008 01:03:00.000615 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sljlg"] Oct 08 01:03:00 crc kubenswrapper[4871]: I1008 01:03:00.324426 4871 generic.go:334] "Generic (PLEG): container finished" podID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerID="2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12" exitCode=0 Oct 08 01:03:00 crc kubenswrapper[4871]: I1008 01:03:00.324518 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sljlg" event={"ID":"f1979d10-0bf0-446c-b6b8-7e6964bbcabb","Type":"ContainerDied","Data":"2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12"} Oct 08 01:03:00 crc kubenswrapper[4871]: I1008 01:03:00.324820 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sljlg" event={"ID":"f1979d10-0bf0-446c-b6b8-7e6964bbcabb","Type":"ContainerStarted","Data":"3c0ca0218181a2a187ba5d209ba027b5dd6dd767c667be3175027e071d4e62fb"} Oct 08 01:03:00 crc kubenswrapper[4871]: I1008 01:03:00.327837 4871 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 01:03:02 crc kubenswrapper[4871]: I1008 01:03:02.344721 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sljlg" event={"ID":"f1979d10-0bf0-446c-b6b8-7e6964bbcabb","Type":"ContainerStarted","Data":"3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d"} Oct 08 01:03:05 crc kubenswrapper[4871]: I1008 01:03:05.379118 4871 generic.go:334] "Generic (PLEG): container finished" podID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerID="3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d" exitCode=0 Oct 08 01:03:05 crc kubenswrapper[4871]: I1008 01:03:05.379761 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sljlg" event={"ID":"f1979d10-0bf0-446c-b6b8-7e6964bbcabb","Type":"ContainerDied","Data":"3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d"} Oct 08 01:03:06 crc kubenswrapper[4871]: I1008 01:03:06.398219 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sljlg" event={"ID":"f1979d10-0bf0-446c-b6b8-7e6964bbcabb","Type":"ContainerStarted","Data":"427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d"} Oct 08 01:03:06 crc kubenswrapper[4871]: I1008 01:03:06.430865 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sljlg" podStartSLOduration=1.9285011779999999 podStartE2EDuration="7.430843303s" podCreationTimestamp="2025-10-08 01:02:59 +0000 UTC" firstStartedPulling="2025-10-08 01:03:00.327623384 +0000 UTC m=+10454.130321457" lastFinishedPulling="2025-10-08 01:03:05.829965469 +0000 UTC m=+10459.632663582" observedRunningTime="2025-10-08 01:03:06.421188049 +0000 UTC m=+10460.223886162" watchObservedRunningTime="2025-10-08 01:03:06.430843303 +0000 UTC m=+10460.233541376" Oct 08 01:03:09 crc kubenswrapper[4871]: I1008 01:03:09.498177 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:03:09 crc kubenswrapper[4871]: I1008 01:03:09.499617 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:03:10 crc kubenswrapper[4871]: I1008 01:03:10.549866 4871 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sljlg" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerName="registry-server" probeResult="failure" output=< Oct 08 01:03:10 crc kubenswrapper[4871]: timeout: failed to connect service ":50051" within 1s Oct 08 01:03:10 crc kubenswrapper[4871]: > Oct 08 01:03:19 crc kubenswrapper[4871]: I1008 01:03:19.592829 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:03:19 crc kubenswrapper[4871]: I1008 01:03:19.680720 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:03:19 crc kubenswrapper[4871]: I1008 01:03:19.841888 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sljlg"] Oct 08 01:03:21 crc kubenswrapper[4871]: I1008 01:03:21.620181 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sljlg" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerName="registry-server" containerID="cri-o://427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d" gracePeriod=2 Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.425327 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.581225 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-utilities\") pod \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.581843 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-catalog-content\") pod \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.582225 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gclm5\" (UniqueName: \"kubernetes.io/projected/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-kube-api-access-gclm5\") pod \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\" (UID: \"f1979d10-0bf0-446c-b6b8-7e6964bbcabb\") " Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.582377 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-utilities" (OuterVolumeSpecName: "utilities") pod "f1979d10-0bf0-446c-b6b8-7e6964bbcabb" (UID: "f1979d10-0bf0-446c-b6b8-7e6964bbcabb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.583266 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.589374 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-kube-api-access-gclm5" (OuterVolumeSpecName: "kube-api-access-gclm5") pod "f1979d10-0bf0-446c-b6b8-7e6964bbcabb" (UID: "f1979d10-0bf0-446c-b6b8-7e6964bbcabb"). InnerVolumeSpecName "kube-api-access-gclm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.644934 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sljlg" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.645098 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sljlg" event={"ID":"f1979d10-0bf0-446c-b6b8-7e6964bbcabb","Type":"ContainerDied","Data":"427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d"} Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.645174 4871 scope.go:117] "RemoveContainer" containerID="427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.644865 4871 generic.go:334] "Generic (PLEG): container finished" podID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerID="427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d" exitCode=0 Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.645847 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sljlg" event={"ID":"f1979d10-0bf0-446c-b6b8-7e6964bbcabb","Type":"ContainerDied","Data":"3c0ca0218181a2a187ba5d209ba027b5dd6dd767c667be3175027e071d4e62fb"} Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.682232 4871 scope.go:117] "RemoveContainer" containerID="3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.684566 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gclm5\" (UniqueName: \"kubernetes.io/projected/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-kube-api-access-gclm5\") on node \"crc\" DevicePath \"\"" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.715529 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1979d10-0bf0-446c-b6b8-7e6964bbcabb" (UID: "f1979d10-0bf0-446c-b6b8-7e6964bbcabb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.725200 4871 scope.go:117] "RemoveContainer" containerID="2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.769960 4871 scope.go:117] "RemoveContainer" containerID="427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d" Oct 08 01:03:22 crc kubenswrapper[4871]: E1008 01:03:22.770619 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d\": container with ID starting with 427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d not found: ID does not exist" containerID="427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.770665 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d"} err="failed to get container status \"427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d\": rpc error: code = NotFound desc = could not find container \"427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d\": container with ID starting with 427097f218af266ab84d1e3d40ffe875eca506456dd2ea6af2822972e300a81d not found: ID does not exist" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.770690 4871 scope.go:117] "RemoveContainer" containerID="3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d" Oct 08 01:03:22 crc kubenswrapper[4871]: E1008 01:03:22.771069 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d\": container with ID starting with 3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d not found: ID does not exist" containerID="3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.771104 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d"} err="failed to get container status \"3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d\": rpc error: code = NotFound desc = could not find container \"3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d\": container with ID starting with 3bef9dcfb5075b9089636a7f21a0e87cb2ea1783823f3ee0ceb7df6ce6cc819d not found: ID does not exist" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.771131 4871 scope.go:117] "RemoveContainer" containerID="2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12" Oct 08 01:03:22 crc kubenswrapper[4871]: E1008 01:03:22.771524 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12\": container with ID starting with 2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12 not found: ID does not exist" containerID="2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.771542 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12"} err="failed to get container status \"2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12\": rpc error: code = NotFound desc = could not find container \"2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12\": container with ID starting with 2b83b416efd3f63bf1bbbab8a2aee66f13022186a1baa8d31736ad4782b8cb12 not found: ID does not exist" Oct 08 01:03:22 crc kubenswrapper[4871]: I1008 01:03:22.786608 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1979d10-0bf0-446c-b6b8-7e6964bbcabb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 01:03:23 crc kubenswrapper[4871]: I1008 01:03:23.017473 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sljlg"] Oct 08 01:03:23 crc kubenswrapper[4871]: I1008 01:03:23.017518 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sljlg"] Oct 08 01:03:25 crc kubenswrapper[4871]: I1008 01:03:25.021865 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" path="/var/lib/kubelet/pods/f1979d10-0bf0-446c-b6b8-7e6964bbcabb/volumes" Oct 08 01:03:35 crc kubenswrapper[4871]: I1008 01:03:35.512811 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 01:03:35 crc kubenswrapper[4871]: I1008 01:03:35.513537 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.426697 4871 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vqzqp"] Oct 08 01:03:52 crc kubenswrapper[4871]: E1008 01:03:52.428193 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerName="extract-content" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.428219 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerName="extract-content" Oct 08 01:03:52 crc kubenswrapper[4871]: E1008 01:03:52.428267 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerName="extract-utilities" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.428279 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerName="extract-utilities" Oct 08 01:03:52 crc kubenswrapper[4871]: E1008 01:03:52.428331 4871 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerName="registry-server" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.428343 4871 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerName="registry-server" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.428706 4871 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1979d10-0bf0-446c-b6b8-7e6964bbcabb" containerName="registry-server" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.433402 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.467096 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqzqp"] Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.566023 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-utilities\") pod \"community-operators-vqzqp\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.566118 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxlhx\" (UniqueName: \"kubernetes.io/projected/88af2d6c-8c82-460d-96b8-7f496cc29a04-kube-api-access-jxlhx\") pod \"community-operators-vqzqp\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.566417 4871 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-catalog-content\") pod \"community-operators-vqzqp\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.669436 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-utilities\") pod \"community-operators-vqzqp\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.669546 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxlhx\" (UniqueName: \"kubernetes.io/projected/88af2d6c-8c82-460d-96b8-7f496cc29a04-kube-api-access-jxlhx\") pod \"community-operators-vqzqp\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.669650 4871 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-catalog-content\") pod \"community-operators-vqzqp\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.669955 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-utilities\") pod \"community-operators-vqzqp\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.670264 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-catalog-content\") pod \"community-operators-vqzqp\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.700854 4871 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxlhx\" (UniqueName: \"kubernetes.io/projected/88af2d6c-8c82-460d-96b8-7f496cc29a04-kube-api-access-jxlhx\") pod \"community-operators-vqzqp\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:52 crc kubenswrapper[4871]: I1008 01:03:52.792205 4871 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:03:53 crc kubenswrapper[4871]: I1008 01:03:53.380854 4871 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqzqp"] Oct 08 01:03:54 crc kubenswrapper[4871]: I1008 01:03:54.142431 4871 generic.go:334] "Generic (PLEG): container finished" podID="88af2d6c-8c82-460d-96b8-7f496cc29a04" containerID="bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29" exitCode=0 Oct 08 01:03:54 crc kubenswrapper[4871]: I1008 01:03:54.142912 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqzqp" event={"ID":"88af2d6c-8c82-460d-96b8-7f496cc29a04","Type":"ContainerDied","Data":"bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29"} Oct 08 01:03:54 crc kubenswrapper[4871]: I1008 01:03:54.142961 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqzqp" event={"ID":"88af2d6c-8c82-460d-96b8-7f496cc29a04","Type":"ContainerStarted","Data":"26a28b2b18c612b0e2aab3f4925bebac557faeed719b4952321119e802023015"} Oct 08 01:03:55 crc kubenswrapper[4871]: I1008 01:03:55.158401 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqzqp" event={"ID":"88af2d6c-8c82-460d-96b8-7f496cc29a04","Type":"ContainerStarted","Data":"1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c"} Oct 08 01:03:56 crc kubenswrapper[4871]: I1008 01:03:56.176896 4871 generic.go:334] "Generic (PLEG): container finished" podID="88af2d6c-8c82-460d-96b8-7f496cc29a04" containerID="1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c" exitCode=0 Oct 08 01:03:56 crc kubenswrapper[4871]: I1008 01:03:56.177434 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqzqp" event={"ID":"88af2d6c-8c82-460d-96b8-7f496cc29a04","Type":"ContainerDied","Data":"1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c"} Oct 08 01:03:57 crc kubenswrapper[4871]: I1008 01:03:57.188644 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqzqp" event={"ID":"88af2d6c-8c82-460d-96b8-7f496cc29a04","Type":"ContainerStarted","Data":"1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7"} Oct 08 01:03:57 crc kubenswrapper[4871]: I1008 01:03:57.214633 4871 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vqzqp" podStartSLOduration=2.6330149130000002 podStartE2EDuration="5.214613038s" podCreationTimestamp="2025-10-08 01:03:52 +0000 UTC" firstStartedPulling="2025-10-08 01:03:54.144911193 +0000 UTC m=+10507.947609266" lastFinishedPulling="2025-10-08 01:03:56.726509318 +0000 UTC m=+10510.529207391" observedRunningTime="2025-10-08 01:03:57.20559467 +0000 UTC m=+10511.008292733" watchObservedRunningTime="2025-10-08 01:03:57.214613038 +0000 UTC m=+10511.017311111" Oct 08 01:04:02 crc kubenswrapper[4871]: I1008 01:04:02.792749 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:04:02 crc kubenswrapper[4871]: I1008 01:04:02.793527 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:04:03 crc kubenswrapper[4871]: I1008 01:04:03.622322 4871 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:04:03 crc kubenswrapper[4871]: I1008 01:04:03.687786 4871 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:04:03 crc kubenswrapper[4871]: I1008 01:04:03.861643 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqzqp"] Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.270258 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vqzqp" podUID="88af2d6c-8c82-460d-96b8-7f496cc29a04" containerName="registry-server" containerID="cri-o://1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7" gracePeriod=2 Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.512296 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.512690 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.804752 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.908923 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxlhx\" (UniqueName: \"kubernetes.io/projected/88af2d6c-8c82-460d-96b8-7f496cc29a04-kube-api-access-jxlhx\") pod \"88af2d6c-8c82-460d-96b8-7f496cc29a04\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.909149 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-catalog-content\") pod \"88af2d6c-8c82-460d-96b8-7f496cc29a04\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.909184 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-utilities\") pod \"88af2d6c-8c82-460d-96b8-7f496cc29a04\" (UID: \"88af2d6c-8c82-460d-96b8-7f496cc29a04\") " Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.910967 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-utilities" (OuterVolumeSpecName: "utilities") pod "88af2d6c-8c82-460d-96b8-7f496cc29a04" (UID: "88af2d6c-8c82-460d-96b8-7f496cc29a04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.927117 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88af2d6c-8c82-460d-96b8-7f496cc29a04-kube-api-access-jxlhx" (OuterVolumeSpecName: "kube-api-access-jxlhx") pod "88af2d6c-8c82-460d-96b8-7f496cc29a04" (UID: "88af2d6c-8c82-460d-96b8-7f496cc29a04"). InnerVolumeSpecName "kube-api-access-jxlhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 01:04:05 crc kubenswrapper[4871]: I1008 01:04:05.961075 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88af2d6c-8c82-460d-96b8-7f496cc29a04" (UID: "88af2d6c-8c82-460d-96b8-7f496cc29a04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.012020 4871 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.012051 4871 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88af2d6c-8c82-460d-96b8-7f496cc29a04-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.012063 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxlhx\" (UniqueName: \"kubernetes.io/projected/88af2d6c-8c82-460d-96b8-7f496cc29a04-kube-api-access-jxlhx\") on node \"crc\" DevicePath \"\"" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.289448 4871 generic.go:334] "Generic (PLEG): container finished" podID="88af2d6c-8c82-460d-96b8-7f496cc29a04" containerID="1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7" exitCode=0 Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.289522 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqzqp" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.289572 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqzqp" event={"ID":"88af2d6c-8c82-460d-96b8-7f496cc29a04","Type":"ContainerDied","Data":"1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7"} Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.289611 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqzqp" event={"ID":"88af2d6c-8c82-460d-96b8-7f496cc29a04","Type":"ContainerDied","Data":"26a28b2b18c612b0e2aab3f4925bebac557faeed719b4952321119e802023015"} Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.289628 4871 scope.go:117] "RemoveContainer" containerID="1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.338551 4871 scope.go:117] "RemoveContainer" containerID="1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.353362 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqzqp"] Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.366208 4871 scope.go:117] "RemoveContainer" containerID="bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.377885 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vqzqp"] Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.410853 4871 scope.go:117] "RemoveContainer" containerID="1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7" Oct 08 01:04:06 crc kubenswrapper[4871]: E1008 01:04:06.411416 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7\": container with ID starting with 1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7 not found: ID does not exist" containerID="1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.411466 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7"} err="failed to get container status \"1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7\": rpc error: code = NotFound desc = could not find container \"1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7\": container with ID starting with 1ad620852214b41f17d359d8952d59994356a0e53743bc219478dfd46c1cdce7 not found: ID does not exist" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.411492 4871 scope.go:117] "RemoveContainer" containerID="1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c" Oct 08 01:04:06 crc kubenswrapper[4871]: E1008 01:04:06.411813 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c\": container with ID starting with 1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c not found: ID does not exist" containerID="1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.411971 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c"} err="failed to get container status \"1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c\": rpc error: code = NotFound desc = could not find container \"1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c\": container with ID starting with 1a58c32c6bb6ca482a6a481ddb485e2c10717f5a8864f04edb3a60333051447c not found: ID does not exist" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.412112 4871 scope.go:117] "RemoveContainer" containerID="bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29" Oct 08 01:04:06 crc kubenswrapper[4871]: E1008 01:04:06.412451 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29\": container with ID starting with bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29 not found: ID does not exist" containerID="bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.412600 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29"} err="failed to get container status \"bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29\": rpc error: code = NotFound desc = could not find container \"bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29\": container with ID starting with bd2283e8d5676fdac84cd6935bf4a4f8807911c428c12ca4ea26c53390d48a29 not found: ID does not exist" Oct 08 01:04:06 crc kubenswrapper[4871]: I1008 01:04:06.993585 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88af2d6c-8c82-460d-96b8-7f496cc29a04" path="/var/lib/kubelet/pods/88af2d6c-8c82-460d-96b8-7f496cc29a04/volumes" Oct 08 01:04:35 crc kubenswrapper[4871]: I1008 01:04:35.512547 4871 patch_prober.go:28] interesting pod/machine-config-daemon-hp4jl container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 01:04:35 crc kubenswrapper[4871]: I1008 01:04:35.514462 4871 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 01:04:35 crc kubenswrapper[4871]: I1008 01:04:35.514566 4871 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" Oct 08 01:04:35 crc kubenswrapper[4871]: I1008 01:04:35.515404 4871 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523"} pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 01:04:35 crc kubenswrapper[4871]: I1008 01:04:35.515560 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerName="machine-config-daemon" containerID="cri-o://07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" gracePeriod=600 Oct 08 01:04:35 crc kubenswrapper[4871]: E1008 01:04:35.644661 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:04:35 crc kubenswrapper[4871]: I1008 01:04:35.737091 4871 generic.go:334] "Generic (PLEG): container finished" podID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" exitCode=0 Oct 08 01:04:35 crc kubenswrapper[4871]: I1008 01:04:35.737134 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" event={"ID":"ef3b09cf-b090-4f2c-892d-ab7f7aee3129","Type":"ContainerDied","Data":"07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523"} Oct 08 01:04:35 crc kubenswrapper[4871]: I1008 01:04:35.737171 4871 scope.go:117] "RemoveContainer" containerID="6b5bdcf7675888de17bfd75524c8b30d3e8713d1d2a3c5cd0e1017c4ecc59890" Oct 08 01:04:35 crc kubenswrapper[4871]: I1008 01:04:35.738148 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:04:35 crc kubenswrapper[4871]: E1008 01:04:35.738441 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:04:48 crc kubenswrapper[4871]: I1008 01:04:48.983185 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:04:48 crc kubenswrapper[4871]: E1008 01:04:48.984272 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:05:00 crc kubenswrapper[4871]: I1008 01:05:00.984903 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:05:00 crc kubenswrapper[4871]: E1008 01:05:00.986118 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:05:13 crc kubenswrapper[4871]: I1008 01:05:13.986191 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:05:13 crc kubenswrapper[4871]: E1008 01:05:13.987129 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:05:19 crc kubenswrapper[4871]: I1008 01:05:19.278732 4871 generic.go:334] "Generic (PLEG): container finished" podID="5ab81d82-f9f2-41f2-adeb-eab439259a96" containerID="49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3" exitCode=0 Oct 08 01:05:19 crc kubenswrapper[4871]: I1008 01:05:19.278868 4871 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b944m/must-gather-xbv7v" event={"ID":"5ab81d82-f9f2-41f2-adeb-eab439259a96","Type":"ContainerDied","Data":"49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3"} Oct 08 01:05:19 crc kubenswrapper[4871]: I1008 01:05:19.280690 4871 scope.go:117] "RemoveContainer" containerID="49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3" Oct 08 01:05:20 crc kubenswrapper[4871]: I1008 01:05:20.129010 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b944m_must-gather-xbv7v_5ab81d82-f9f2-41f2-adeb-eab439259a96/gather/0.log" Oct 08 01:05:25 crc kubenswrapper[4871]: I1008 01:05:25.982068 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:05:25 crc kubenswrapper[4871]: E1008 01:05:25.982829 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:05:29 crc kubenswrapper[4871]: I1008 01:05:29.694338 4871 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b944m/must-gather-xbv7v"] Oct 08 01:05:29 crc kubenswrapper[4871]: I1008 01:05:29.695463 4871 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-b944m/must-gather-xbv7v" podUID="5ab81d82-f9f2-41f2-adeb-eab439259a96" containerName="copy" containerID="cri-o://b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2" gracePeriod=2 Oct 08 01:05:29 crc kubenswrapper[4871]: I1008 01:05:29.705951 4871 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b944m/must-gather-xbv7v"] Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.214637 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b944m_must-gather-xbv7v_5ab81d82-f9f2-41f2-adeb-eab439259a96/copy/0.log" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.215632 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.391726 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5ab81d82-f9f2-41f2-adeb-eab439259a96-must-gather-output\") pod \"5ab81d82-f9f2-41f2-adeb-eab439259a96\" (UID: \"5ab81d82-f9f2-41f2-adeb-eab439259a96\") " Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.392238 4871 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrljb\" (UniqueName: \"kubernetes.io/projected/5ab81d82-f9f2-41f2-adeb-eab439259a96-kube-api-access-wrljb\") pod \"5ab81d82-f9f2-41f2-adeb-eab439259a96\" (UID: \"5ab81d82-f9f2-41f2-adeb-eab439259a96\") " Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.404115 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab81d82-f9f2-41f2-adeb-eab439259a96-kube-api-access-wrljb" (OuterVolumeSpecName: "kube-api-access-wrljb") pod "5ab81d82-f9f2-41f2-adeb-eab439259a96" (UID: "5ab81d82-f9f2-41f2-adeb-eab439259a96"). InnerVolumeSpecName "kube-api-access-wrljb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.431352 4871 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b944m_must-gather-xbv7v_5ab81d82-f9f2-41f2-adeb-eab439259a96/copy/0.log" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.431941 4871 generic.go:334] "Generic (PLEG): container finished" podID="5ab81d82-f9f2-41f2-adeb-eab439259a96" containerID="b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2" exitCode=143 Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.431986 4871 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b944m/must-gather-xbv7v" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.432000 4871 scope.go:117] "RemoveContainer" containerID="b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.478486 4871 scope.go:117] "RemoveContainer" containerID="49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.495382 4871 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrljb\" (UniqueName: \"kubernetes.io/projected/5ab81d82-f9f2-41f2-adeb-eab439259a96-kube-api-access-wrljb\") on node \"crc\" DevicePath \"\"" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.529751 4871 scope.go:117] "RemoveContainer" containerID="b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2" Oct 08 01:05:30 crc kubenswrapper[4871]: E1008 01:05:30.530283 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2\": container with ID starting with b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2 not found: ID does not exist" containerID="b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.530336 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2"} err="failed to get container status \"b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2\": rpc error: code = NotFound desc = could not find container \"b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2\": container with ID starting with b995a03c4b65aabf2ed14fc07d8e30d90529c6d85af9f146beff7557610f8ff2 not found: ID does not exist" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.530370 4871 scope.go:117] "RemoveContainer" containerID="49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3" Oct 08 01:05:30 crc kubenswrapper[4871]: E1008 01:05:30.530673 4871 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3\": container with ID starting with 49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3 not found: ID does not exist" containerID="49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.530715 4871 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3"} err="failed to get container status \"49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3\": rpc error: code = NotFound desc = could not find container \"49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3\": container with ID starting with 49e20e1ef7594c45eab2d560441609052739b2ca522dc840b5627410dfb78bb3 not found: ID does not exist" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.594950 4871 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ab81d82-f9f2-41f2-adeb-eab439259a96-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5ab81d82-f9f2-41f2-adeb-eab439259a96" (UID: "5ab81d82-f9f2-41f2-adeb-eab439259a96"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.597333 4871 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5ab81d82-f9f2-41f2-adeb-eab439259a96-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 01:05:30 crc kubenswrapper[4871]: I1008 01:05:30.997970 4871 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ab81d82-f9f2-41f2-adeb-eab439259a96" path="/var/lib/kubelet/pods/5ab81d82-f9f2-41f2-adeb-eab439259a96/volumes" Oct 08 01:05:40 crc kubenswrapper[4871]: I1008 01:05:40.983256 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:05:40 crc kubenswrapper[4871]: E1008 01:05:40.984115 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:05:54 crc kubenswrapper[4871]: I1008 01:05:54.982885 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:05:54 crc kubenswrapper[4871]: E1008 01:05:54.983896 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:06:04 crc kubenswrapper[4871]: I1008 01:06:04.969522 4871 scope.go:117] "RemoveContainer" containerID="06804e71303821f5dd85704929191cb0e1d15fd6478514154104cc7fcfbce9bc" Oct 08 01:06:06 crc kubenswrapper[4871]: I1008 01:06:06.996380 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:06:06 crc kubenswrapper[4871]: E1008 01:06:06.997727 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:06:18 crc kubenswrapper[4871]: I1008 01:06:18.983198 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:06:18 crc kubenswrapper[4871]: E1008 01:06:18.986335 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:06:30 crc kubenswrapper[4871]: I1008 01:06:30.982950 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:06:30 crc kubenswrapper[4871]: E1008 01:06:30.984067 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:06:42 crc kubenswrapper[4871]: I1008 01:06:42.983828 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:06:42 crc kubenswrapper[4871]: E1008 01:06:42.985842 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:06:55 crc kubenswrapper[4871]: I1008 01:06:55.982397 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:06:55 crc kubenswrapper[4871]: E1008 01:06:55.983695 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" Oct 08 01:07:07 crc kubenswrapper[4871]: I1008 01:07:07.983036 4871 scope.go:117] "RemoveContainer" containerID="07c2a2eff927879ee2cb34312cfca8025d359434420a218ce182f44e1af56523" Oct 08 01:07:07 crc kubenswrapper[4871]: E1008 01:07:07.988659 4871 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hp4jl_openshift-machine-config-operator(ef3b09cf-b090-4f2c-892d-ab7f7aee3129)\"" pod="openshift-machine-config-operator/machine-config-daemon-hp4jl" podUID="ef3b09cf-b090-4f2c-892d-ab7f7aee3129" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071334516024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071334517017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071307225016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071307225015457 5ustar corecore